Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Deep neural network hardware accelerator based on power exponent quantization

A technology of deep neural network and hardware accelerator, applied in the direction of biological neural network model, neural architecture, neural learning method, etc., can solve the problems of complex processor circuit, high power consumption, huge storage space, etc., to improve computing speed, high Calculation performance and the effect of reducing the amount of calculation

Active Publication Date: 2019-10-29
SOUTHEAST UNIV
View PDF3 Cites 11 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the invention of the present invention is to aim at the deficiency of above-mentioned background technology, provide a kind of deep neural network hardware accelerator based on exponentiation exponent quantization, by designing the hardware accelerator that realizes shift operation based on the result of exponent exponent quantization to neural network parameter, avoid Using complex multiplication circuits to realize floating-point multiplication operations reduces processor power consumption and chip area, and solves the technical problems of complex circuits, large storage space and high power consumption in existing processors that implement deep neural network convolution operations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Deep neural network hardware accelerator based on power exponent quantization
  • Deep neural network hardware accelerator based on power exponent quantization
  • Deep neural network hardware accelerator based on power exponent quantization

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] The technical solution of the invention will be described in detail below in conjunction with the accompanying drawings.

[0024] The deep neural network accelerator hardware structure that the present invention designs is as figure 1 As shown, the PE array size is 16*16, the convolution kernel size is 3*3, and the convolution kernel step size is 1 as an example to illustrate its working method: the accelerator quantizes the input data and the power exponent weight data through the AXI-4 bus Cache to the input buffer area and the weight buffer area. According to the size of the PE array, the AXI-4 bus needs to read 16 convolution kernel data from the DDR and store them in the weight buffer area, so as to input 16 weight index values ​​​​to the PE array in parallel. The convolution kernel data stored in the weight buffer is read to the encoding module, and then the weight data is encoded in a certain way according to the positive or negative of the weight data quantized ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a deep neural network hardware accelerator based on power exponent quantization, relates to a processor structure for deep neural network convolution calculation hardware acceleration, and belongs to the technical field of calculation, calculation and counting. The hardware accelerator comprising: an AXI-4 bus interface, an input cache region, an output cache region, a weight cache region, a weight index cache region, an encoding module, a configurable state controller module and a PE array. The input cache region and the output cache region are designed into a row cache structure; the encoder encodes the weights according to an ordered quantization set that stores absolute values of all quantized weights. when the accelerator performs calculation, the PE unit readsdata from the input cache region and the weight index cache region to perform shift calculation, and sends a calculation result to the output cache region. According to the method, floating point multiplication is replaced by shift operation, so that the requirements on computing resources, storage resources and communication bandwidth are reduced, and the computing efficiency of the acceleratoris further improved.

Description

technical field [0001] The invention discloses a deep neural network hardware accelerator based on power exponent quantization, relates to a processor structure for hardware acceleration of deep neural network convolution calculation, and belongs to the technical field of calculation, calculation and counting. Background technique [0002] In recent years, artificial intelligence has penetrated into all areas of life and has had a major impact on the world's economic and social activities. Deep learning is a hot field in machine learning research and is widely used in mainstream artificial intelligence algorithms. As a technology of deep learning, Deep Convolutional Neural Network (DCNN) is currently widely used in many Artificial Intelligence (AI) applications, deep convolutional neural networks have achieved some remarkable results in the technical fields of computer vision, speech recognition and robotics, especially in the field of image recognition. [0003] Deep convo...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/063G06N3/08
CPCG06N3/063G06N3/08G06N3/045G06F2207/4824G06F5/01G06F7/552G06F7/556G06F9/5027
Inventor 陆生礼庞伟武瑞利樊迎博刘昊黄成
Owner SOUTHEAST UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products