Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-precision neural network computing device and method based on data stream architecture

A neural network and computing device technology, applied in the computer field, can solve problems such as waste of operating bandwidth, waste of power consumption, and lack of instruction support, and achieve the effects of reducing computing delay, avoiding data overflow, and improving user experience

Active Publication Date: 2021-08-24
INST OF COMPUTING TECHNOLOGY - CHINESE ACAD OF SCI
View PDF6 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, on the existing accelerators, since there are no relevant computing components to support low-precision calculations, and there is no corresponding instruction support, these low-precision calculations cannot be supported
Therefore, low-precision network calculations cannot be performed to reduce the power consumption of the accelerator
If the original precision operation unit is used to perform low-precision calculations, the power consumption is wasted and there is no performance improvement
In addition, the results of low-precision calculations are usually saved with a wide bit width to avoid data overflow. In addition, the use of SIMD operations in existing acceleration devices has high data bandwidth (on-chip network bandwidth). When the designed low-precision components are unreasonable, Not only will the operating bandwidth be wasted, but it will also cause data overflow, which makes it challenging to support multiple low-precision component designs

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-precision neural network computing device and method based on data stream architecture
  • Multi-precision neural network computing device and method based on data stream architecture
  • Multi-precision neural network computing device and method based on data stream architecture

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0032] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below through specific embodiments in conjunction with the accompanying drawings. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0033] As mentioned in the background technology section, the use of SIMD operations in existing computing devices has high data bandwidth (on-chip network bandwidth). When the designed low-precision components are unreasonable, not only will the operating bandwidth be wasted, but also data of overflow. Therefore, the present application constructs a multi-precision neural network computing device based on a data flow architecture, including: a microcontroller and a PE array connected to it, each PE of the PE array is configured with an original precision and a precision lower than the...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention provides a multi-precision neural network computing device based on a data stream architecture, which comprises a microcontroller and a PE array connected with the microcontroller, and each PE of the PE array is configured with a plurality of low-precision computing components with original precision and precision lower than the original precision. The computing component with lower precision is provided with more parallel multiply accumulators to make full use of the on-chip network bandwidth, and each low-precision computing component in each PE is provided with sufficient registers to avoid data overflow; and the microcontroller is configured to respond to an acceleration request for a specific convolutional neural network, control a calculation component with original precision or low precision matched with the precision of the specific convolutional neural network in the PE array to execute operation in corresponding convolution operation and store an intermediate calculation result to a corresponding register. Therefore, the convolutional neural networks with different precisions can be accelerated, the calculation time delay and energy consumption are reduced, and the user experience is improved.

Description

technical field [0001] The present invention relates to the field of computers, in particular to the field of acceleration devices or accelerators for neural network model calculations, and more specifically, to a multi-precision neural network calculation device and method based on a data flow architecture. Background technique [0002] Deep Neural Network (DNN) has shown significant advantages in many application fields, from computer vision to natural language processing and speech recognition. The support of powerful hardware makes it easier to train DNN models. With the diversification and complexity of applications, The DNN model is also complex, its depth is getting deeper and its parameters are getting larger and larger, and the complex DNN model is more expressive in capturing the nonlinear relationship between features and output, thus showing good result accuracy. [0003] Although these neural networks are very powerful, a large number of weights will occupy a la...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/063G06N3/04
CPCG06N3/063G06N3/045
Inventor 吴欣欣范志华欧焱李文明叶笑春范东睿
Owner INST OF COMPUTING TECHNOLOGY - CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products