Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network model compression method, system and device and medium

A technology of neural network model and compression method, which is applied in the field of model compression based on meta-learning and soft pruning, and can solve the problems that cannot be changed according to the situation

Inactive Publication Date: 2020-04-28
INFORMATION SCI RES INST OF CETC +1
View PDF0 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0010] Once the compression strategy is selected, the selected strategy will remain constant throughout the compression debugging process and cannot be adjusted according to changes in the situation, such as changes in the probability distribution of filter parameters or changes in the architecture of the deep learning model

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network model compression method, system and device and medium
  • Neural network model compression method, system and device and medium
  • Neural network model compression method, system and device and medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0053] In order to enable those skilled in the art to better understand the technical solutions of the present invention, the present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments.

[0054] Based on filter pruning, it is necessary to select a compression strategy in advance based on experience, such as a strategy based on weight size and a strategy based on the similarity between filters. Once the compression strategy is selected, the selected strategy will remain constant throughout the compression debugging process and cannot be adjusted according to changes in the situation, such as changes in the probability distribution of filter parameters or changes in the deep learning model architecture. The present invention provides a neural network model compression method based on meta-learning, which proposes the concept and scheme of meta-learning pruning on the basis of soft pruning filtering, so as to ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a compression method and system for a neural network model, electronic equipment and a readable medium. The method comprises the steps: presetting a pruning strategy of each network layer, and the pruning strategy comprises a pruning rate, a pruning threshold and a filter weight value; in the current compression period, selecting a pruning filter to be pruned from each network layer according to a pruning strategy, and setting the value of each pruning filter to be zero; setting the value of the pruning filter to be a non-zero value through fine adjustment so as to update the neural network model and obtain a current neural network model; and determining whether the pruning strategy of the next compression period needs to be adjusted or not according to the current meta-attribute value output by the current neural network model. Allowing the pruned filter to update without reducing the number of feature maps of the network; an extra model fine adjustment stage isnot needed any more, and the network training time is shortened; according to the parameter statistical distribution of the current model, the pruning strategy most suitable for the current model parameters is selected, and the model training effect is improved.

Description

technical field [0001] The invention belongs to the technical field of neural network compression, and in particular relates to a model compression method, in particular to a model compression method based on meta-learning and soft pruning. Background technique [0002] The mainstream methods for compressing and accelerating neural networks can be divided into five types: 1) parameter pruning; 2) parameter sharing; 3) low-rank decomposition; 4) compact convolution Core design (designing compact convolutional filters); 5) knowledge distillation. Parameter pruning mainly removes redundant parameters by designing criteria for judging whether parameters are important or not. Parameter sharing mainly explores the redundancy of model parameters, and uses technologies such as Hash or quantization to compress weights. Low-rank decomposition utilizes matrix or tensor decomposition techniques to estimate and decompose the original convolution kernels in deep models. The design of t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/082G06N3/045
Inventor 陈文彬王子玮张峰胡金晖
Owner INFORMATION SCI RES INST OF CETC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products