Hardware accelerator applied to binarized convolutional neural network and data processing method thereof

A technology of convolutional neural network and hardware accelerator, which is applied in neural learning methods, electrical digital data processing, biological neural network models, etc., and can solve problems such as increased computation, high resource consumption, and increased computing cycles

Active Publication Date: 2020-02-11
HEFEI UNIV OF TECH
View PDF11 Cites 11 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, among the many technologies they adopt, the feature map input to the first layer still uses full-precision data, and a special computing unit needs to be designed for the first layer calculation, which has poor versatility, high resource consumption and large power loss; traditional Convolution adopts the strategy of multiply-accumulate and popcount algorithm, which does not make maximum use of FPGA internal resources, and the calculation cycle increases; at the same time, the strategy of convolution calculation edge complement +1 or -1 is adopted, the edge data still needs to be calculated, and the amount of calculation increases accordingly , leading to greater consumption of hardware resources and increased computing cycles

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Hardware accelerator applied to binarized convolutional neural network and data processing method thereof
  • Hardware accelerator applied to binarized convolutional neural network and data processing method thereof
  • Hardware accelerator applied to binarized convolutional neural network and data processing method thereof

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0081] In this embodiment, the binary convolutional neural network includes: a K-layer binary convolution layer, a K-layer activation function layer, a K-batch normalization layer, a K-layer pooling layer, and a fully connected classification output layer; and The number of training parameters in the batch normalization layer is combined into one;

[0082] The convolutional neural network adopted in this embodiment is a handwritten digit recognition network, and the structure diagram is as follows figure 1 As shown, its structure includes an input layer, two convolutional layers, two pooling layers, and a fully connected layer. The first layer calculation is the convolutional layer calculation, the input layer is 784 neural nodes, the convolution kernel size is 3×3, the convolution step is 1, and 16 feature maps of 28×28 are output, with a total of 12544 neural nodes; The second layer is the calculation of the pooling layer. The input layer is 16 28×28 feature maps output by...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a hardware accelerator applied to a binarized convolutional neural network and a data processing method of the hardware accelerator. The hardware accelerator comprises a neuralnetwork parameter storage module, a matrix generator module, a convolution calculation array module, a pooling layer module and a global control module. A binary grayscale picture and neural networktraining parameters of each layer are pre-stored in the neural network parameter storage module; the matrix generator is responsible for preparing input data of convolution operation; the convolutioncalculation array is responsible for convolution calculation of a convolution layer; the pooling layer module is responsible for performing pooling operation on the output of the convolution layer; the global control module is responsible for controlling the whole system to work normally. The hardware accelerator aims to improve the operation speed of the convolutional neural network, reduce resources and computing resources consumed by network deployment on a hardware platform, and reduce the power consumption of network operation at the same time.

Description

technical field [0001] The invention belongs to the field of artificial intelligence hardware design, and in particular relates to an accelerator applied to a binary convolutional neural network and a data processing method thereof. Background technique [0002] Convolutional neural networks are derived from artificial neural networks. As a multi-layer perceptual network, it has strong adaptability to various image transformation forms such as image rotation, scaling down or enlargement, and image translation, and can quickly extract image features. It adopts the weight sharing network structure, which has a strong similarity with the biological neural network structure. This structure reduces the number of weights, thereby reducing the complexity of the network model. When inputting multidimensional images to the network, such The advantages of the algorithm are more obvious, so that the image can be directly used as the input of the network, avoiding the complex feature e...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F9/30G06N3/04G06N3/063G06N3/08
CPCG06N3/08G06N3/063G06F9/30007G06F9/30098G06N3/045
Inventor 杜高明涂振兴陈邦溢杨振文张多利宋宇鲲李桢旻
Owner HEFEI UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products