Memory normalization network layer method based on double forward propagation algorithm

A forward propagation and normalization technology, applied in the design of the network layer and the research field of neural network training algorithms, can solve the problems of feature distribution deviation, inconsistency, model deviation, etc., to improve training stability and reduce training instability. problems, the effect of reducing the problem of inaccurate calculation of statistics

Pending Publication Date: 2020-01-17
SOUTH CHINA UNIV OF TECH
View PDF2 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In other words, the training process of BN is inconsistent with the testing process, resulting in model deviation
In particular, each stochastic gradient descent (SGD) during the training process will cause a change in the model parameters, which in turn will lead to deviations in the feature distribution before and after the model is updated.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Memory normalization network layer method based on double forward propagation algorithm
  • Memory normalization network layer method based on double forward propagation algorithm
  • Memory normalization network layer method based on double forward propagation algorithm

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0042] A kind of memory normalization network layer method based on the double forward propagation algorithm of the present embodiment 1, such as figure 1 shown, including the following steps:

[0043] S1. Estimate the statistical information of multiple batches in the memory storage and normalize the training data;

[0044] S2. After performing the standard forward and backward propagation, the double forward propagation algorithm performs an additional forward propagation to reduce the distribution deviation of the network middle layer information caused by the update of the neural network model.

[0045] The memory is stored as a real-valued vector with dimension K, which records the statistics of the data distribution recorded in the latest K iterations.

[0046] Such as figure 2 As shown, the distribution of the training samples is estimated according to the statistical information of multiple batches in the memory storage, and the specific calculation method is as fo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a memory normalization network layer method based on a double forward propagation algorithm, so as to improve the training stability and network performance of a neural network. The method comprises the steps that statistical information of each patch is calculated, the patches refer to a set of data samples, and the statistical information comprises the mean value and variance of data; estimating an overall mean value and an overall variance of a plurality of continuous patches in the storage memory according to the statistical magnitude of each patch, and performing normalization operation on each generation of training data in the training process by using the statistical magnitude; after standard forward propagation and back propagation are executed once in eachiteration, forward propagation is additionally executed once to reduce distribution offset of network interlayer information caused by model updating. According to the method, more accurate and morestable statistical information can be obtained, consistent behaviors are always kept in training and testing, and finally, the generalization ability of the deep network can be greatly improved.

Description

technical field [0001] The invention relates to the technical field of deep learning and algorithm optimization, in particular to the design of network layers and the research on training algorithms of neural networks. Background technique [0002] Batch Normalization (BN) has become a standard component in deep networks. The most critical technology is that this method normalizes the data according to the statistics of the current batch data, so that it is distributed in a standard distribution, thus effectively solving the problem of distribution deviation in the training process. By introducing BN into the network structure, the training process of the model can converge faster and obtain better generalization ability. [0003] Although BN plays a vital role in the training of deep networks, it still has some limitations. On the one hand, during the training process, BN estimates the mean and variance of the overall distribution of samples based on the data of a single ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/084G06N3/045
Inventor 谭明奎吴庆耀国雍
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products