Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Artificial neural network calculating device and method for sparse connection

An artificial neural network and computing device technology, applied in neural learning methods, biological neural network models, calculations, etc., can solve problems such as performance bottlenecks, high power consumption overhead, etc., to improve support, avoid performance bottlenecks, and reduce memory access bandwidth Effect

Active Publication Date: 2016-04-20
CAMBRICON TECH CO LTD
View PDF2 Cites 261 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, since the GPU is a device specially used to perform graphics and image operations and scientific computing, it does not have special support for sparse artificial neural network operations, and still requires a lot of front-end decoding work to perform sparse artificial neural network operations, which brings a lot of problems. additional overhead
In addition, the GPU has only a small on-chip cache, and the model data (weights) of the multi-layer artificial neural network need to be repeatedly moved from off-chip. The off-chip bandwidth has become the main performance bottleneck, and it has brought huge power consumption overhead.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Artificial neural network calculating device and method for sparse connection
  • Artificial neural network calculating device and method for sparse connection
  • Artificial neural network calculating device and method for sparse connection

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0050] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with specific embodiments and with reference to the accompanying drawings.

[0051] The invention discloses an artificial neural network computing device for sparse connections, comprising:

[0052] The mapping unit is used to convert the input data into a one-to-one storage method of input neurons and weights, and store them in a storage device or a cache;

[0053] storage device for storing data and instructions;

[0054] An operation unit, configured to perform corresponding operations on the data according to instructions stored in the storage device; the operation unit mainly performs three-step operations, the first step is to multiply the input neurons and weight data; the second step The second step is to perform the addition tree operation, which is used to add the weighted output neuron...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

An artificial neural network calculating device for sparse connection comprises a mapping unit used for converting input data into the storage mode that input nerve cells and weight values correspond one by one, a storage unit used for storing data and instructions, and an operation unit used for executing corresponding operation on the data according to the instructions. The operation unit mainly executes three steps of operation, wherein in the first step, the input nerve cells and weight value data are multiplied; in the second step, addition tree operation is executed, the weighted output nerve cells processed in the first step are added level by level through an addition tree, or the output nerve cells are added with offset to obtain offset-added output nerve cells; in the third step, activation function operation is executed, and the final output nerve cells are obtained. By means of the device, the problems that the operation performance of a CPU and a GPU is insufficient, and the expenditure of front end coding is large are solved, support to a multi-layer artificial neural network operation algorithm is effectively improved, and the problem that memory bandwidth becomes a bottleneck of multi-layer artificial neural network operation and the performance of a training algorithm of the multi-layer artificial neural network operation is solved.

Description

technical field [0001] The present invention relates to the technical field of data processing, and more particularly to an artificial neural network computing device and method for sparse connections. Background technique [0002] Artificial Neural Networks (ANNs), referred to as Neural Networks (NNs) for short, is an algorithmic mathematical model that imitates the behavioral characteristics of animal neural networks and performs distributed parallel information processing. This kind of network depends on the complexity of the system, and achieves the purpose of processing information by adjusting the interconnection relationship between a large number of internal nodes. The algorithm used in neural networks is vector multiplication, and symbolic functions and their various approximations are widely used. [0003] Just like the neural network in the brain, the neural network consists of some interconnected nodes, such as figure 1 As shown, each circle represents a neuron...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04
CPCG06F9/30007G06F9/3802G06F9/3887G06N3/04G06N3/063G06N3/048G06N3/045G06F2207/4824G06N3/082G06F12/1081G06F2212/621G06F18/214G06F9/30145G06F17/16G06F7/5443G06F17/15
Inventor 张士锦郭崎陈云霁陈天石
Owner CAMBRICON TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products