Hardware Platform-Aware Binarized Neural Network Model Optimization
Open Access
- 26 January 2022
- journal article
- research article
- Published by MDPI AG in Applied Sciences
- Vol. 12 (3), 1296
- https://doi.org/10.3390/app12031296
Abstract
Deep Neural Networks (DNNs) have shown superior accuracy at the expense of high memory and computation requirements. Optimizing DNN models regarding energy and hardware resource requirements is extremely important for applications with resource-constrained embedded environments. Although using binary neural networks (BNNs), one of the recent promising approaches, significantly reduces the design’s complexity, accuracy degradation is inevitable when reducing the precision of parameters and output activations. To balance between implementation cost and accuracy, in addition to proposing specialized hardware accelerators for corresponding specific network models, most recent software binary neural networks have been optimized based on generalized metrics, such as FLOPs or MAC operation requirements. However, with the wide range of hardware available today, independently evaluating software network structures is not good enough to determine the final network model for typical devices. In this paper, an architecture search algorithm based on estimating the hardware performance at the design time is proposed to achieve the best binary neural network models for hardware implementation on target platforms. With the XNOR-net used as a base architecture and target platforms, including Field Programmable Gate Array (FPGA), Graphic Processing Unit (GPU), and Resistive Random Access Memory (RRAM), the proposed algorithm shows its efficiency by giving more accurate estimation for the hardware performance at the design time than FLOPs or MAC operations.This publication has 25 references indexed in Scilit:
- FP-BNN: Binarized neural network on FPGANeurocomputing, 2018
- Exploring Heterogeneous Algorithms for Accelerating Deep Convolutional Neural Networks on FPGAsPublished by Association for Computing Machinery (ACM) ,2017
- A Kernel Decomposition Architecture for Binary-weight Convolutional Neural NetworksPublished by Association for Computing Machinery (ACM) ,2017
- Evaluating Fast Algorithms for Convolutional Neural Networks on FPGAsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2017
- FINNPublished by Association for Computing Machinery (ACM) ,2017
- Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural NetworksIEEE Journal of Solid-State Circuits, 2016
- Neural networks designing neural networksPublished by Association for Computing Machinery (ACM) ,2016
- XNOR-Net: ImageNet Classification Using Binary Convolutional Neural NetworksPublished by Springer Science and Business Media LLC ,2016
- Quantized Convolutional Neural Networks for Mobile DevicesPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2016
- Going Deeper with Embedded FPGA Platform for Convolutional Neural NetworkPublished by Association for Computing Machinery (ACM) ,2016