Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Activation function parameterization improvement method based on recurrent neural network

A technology of cyclic neural network and activation function, applied in neural learning methods, biological neural network models, neural architectures, etc., can solve the problem of increasing the difficulty of cyclic neural network training, activation function falling into the saturation region, and inability to effectively correct weights, etc. problem, to achieve the effect of preventing gradient disappearance, avoiding too small derivative, and good effect

Pending Publication Date: 2019-06-07
NANJING UNIV OF POSTS & TELECOMM +1
View PDF0 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] In order to solve the deficiencies in the prior art, the present invention provides an activation function parameterization improvement method based on a recurrent neural network, which solves the two-way saturation characteristic of the S-type activation function, which easily leads to the disappearance of the gradient and cannot effectively correct the weight value, and in When performing reverse error propagation, it is easy to cause the activation function to fall into the saturation region, causing the gradient to disappear, which increases the difficulty of training the recurrent neural network.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Activation function parameterization improvement method based on recurrent neural network
  • Activation function parameterization improvement method based on recurrent neural network
  • Activation function parameterization improvement method based on recurrent neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0032] The present invention will be further described below in conjunction with the accompanying drawings. The following examples are only used to illustrate the technical solution of the present invention more clearly, but not to limit the protection scope of the present invention.

[0033] Based on the DC-Bi-LSTM network, the present invention designs a universal Sigmoid activation function form, modifies the activation function expression in combination with the internal structure of LSTM, and confirms the function expression form by testing different parameter combinations. This improves the accuracy of text classification.

[0034] A method for improving parameterization of an activation function based on a recurrent neural network, comprising the following steps:

[0035] In step 1, a bidirectional long-short-term memory network (Bi-LSTM) is constructed on the basis of a long-term short-term memory network (LSTM); in this embodiment, the Bi-LSTM has 15 layers.

[0036...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an activation function parameterization improvement method based on a recurrent neural network, and the method comprises the steps: step 1, constructing a bidirectional long short-term memory network Bi-LSTM on the basis of a long short-term memory network; step 2, connecting all hidden layers in the Bi-LSTM network in series, adding an average pooling layer behind the lasthidden layer in the network, connecting a normalized exponential function layer behind the average pooling layer, and establishing a densely connected bidirectional long short-term memory network DC-Bi-LSTM; and step 3, training on the data set by applying the parameterized Sigmoid activation function, recording the sentence classification accuracy of the densely connected bidirectional long-short-term memory network, and obtaining the parameterized activation function corresponding to the optimal accuracy. According to the invention, through the parameterized activation function module, theunsaturated region of the S-shaped activation function is expanded, the derivative of the function is prevented from being too small, and the gradient disappearance phenomenon is prevented.

Description

technical field [0001] The present invention relates to the technical field of natural language processing and text classification, in particular to an activation function parameterization improvement method based on a recurrent neural network. Background technique [0002] Deep neural networks are widely used in computer vision, however, stacked recurrent neural networks suffer from vanishing gradient and overfitting problems. Therefore, on this basis, some new types of cyclic neural networks have emerged, and the bidirectional long-term short-term memory network based on dense connections is a very effective cyclic neural network. [0003] The activation function module is a basic module of the neural network. In general, activation functions have the following properties: [0004] (1) Nonlinearity: Almost all functions can be expressed by a two-layer neural network under the condition that the activation function is nonlinear; [0005] (2) Differentiability: The optimi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/35G06F17/27G06N3/04G06N3/08
Inventor 于舒娟李润琦高冲杨杰张昀
Owner NANJING UNIV OF POSTS & TELECOMM
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products