Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network model compression method based on sparse backward propagation training

A neural network model and back-propagation technology, applied in biological neural network models, neural learning methods, etc., can solve the problems of long time-consuming training and inference of neural networks, difficult deep neural networks, and low accuracy of neural network models.

Inactive Publication Date: 2018-03-23
PEKING UNIV
View PDF0 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The disadvantage of the method is that the model is often too large. On the one hand, the training and reasoning of the neural network takes a long time, and the amount of calculation and power consumption is large; The accuracy of the network model is low, resulting in overfitting (Overfitting) phenomenon
In view of the problems existing in the above methods, some model compression algorithms have been proposed successively, such as the weight clipping algorithm for feedforward neural network, the weight progressive clipping algorithm for recurrent neural network, and the parameter quantization method for convolutional neural network, etc.; however, , because for any layer in the deep neural network model, the above algorithm needs to specify hyperparameters separately to control the last dimension of the layer, therefore, these existing methods are not flexible in practical applications, and it is difficult to apply to the deep neural network. train

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network model compression method based on sparse backward propagation training
  • Neural network model compression method based on sparse backward propagation training
  • Neural network model compression method based on sparse backward propagation training

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0056] Below in conjunction with accompanying drawing, further describe the present invention through embodiment, but do not limit the scope of the present invention in any way.

[0057] The present invention provides a sparse backpropagation training method for a neural network model, which is a sparse backpropagation training method based on a large value of k. figure 1 It is a block flow diagram of the method of the present invention.

[0058] The most basic computing unit of a neural network is a linear transformation plus a nonlinear transformation. Therefore, the specific implementation takes the most basic neural network computing unit as an example, and the calculation formula is as follows:

[0059] y=Wx

[0060] z=σ(y)

[0061] Among them, W is the parameter matrix of the model, x is the input vector, y is the output after linear transformation, σ is a function of nonlinear transformation, and z is the output after nonlinear transformation.

[0062] The sparse ba...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a sparse backward propagation compression method of a neural network model, belongs to the field of information technology, and relates to machine learning and deep learning technologies. In the process of backward propagation, each layer of the neural network model uses the output gradient of the previous layer as the input to calculate the gradient, and performs k large-value sparse processing to obtain the sparsely processed vector and the number of sparse return times, and record k The index corresponding to the value; use the sparse gradient to update the parameters of the neural network; according to the k-value subscript index, delete the neuron with a small number of return times, and compress the model. The present invention adopts a sparse method based on a large value of k in the backward propagation process, eliminates inactive neurons, compresses the size of the model, improves the training and reasoning speed of the deep neural network, and maintains good precision.

Description

technical field [0001] The invention belongs to the field of information technology, relates to machine learning and deep learning technology, and in particular to a compression algorithm based on a large k-value sparse (Sparsify) neural network model. Background technique [0002] The training process of the neural network model is divided into two steps: forward propagation (Forward Propagation) and backward propagation (Backward Propagation). The forward propagation calculates the output of each layer of the neural network, and the backward propagation calculates the gradient of the model parameters according to the gradient of the output result of the forward propagation, and finally the neural network model uses these gradients to update the parameters. However, in practical applications, the deployed model only performs inference operations, that is, forward propagation operations. [0003] In existing neural network model training methods, the size, or dimension, of ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08
CPCG06N3/08G06N3/084
Inventor 孙栩任宣丞马树铭位冰镇
Owner PEKING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products