Training method and system of deep neural network based on critical damping momentum

A deep neural network and neural network technology, applied in the field of numerical and machine learning, can solve the problems of non-convergence, slow pace of neural network parameters, and few network types, etc., achieve fast convergence speed, and accelerate the process of training and deployment Effect

Pending Publication Date: 2021-03-12
SHENZHEN GRADUATE SCHOOL TSINGHUA UNIV
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the iteration speed of the SGD algorithm is relatively slow, and its learning rate is a hyperparameter that can only be set empirically
When the learning rate is set too small or too large, the iteration of the neural network parameters is prone to the problem of slow pace or violent oscillation.
Although the Adam algorithm can obtain faster speed than SGD under certain conditions, it is applicable to fewer network types than SGD, and may not converge in some special cases.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Training method and system of deep neural network based on critical damping momentum
  • Training method and system of deep neural network based on critical damping momentum
  • Training method and system of deep neural network based on critical damping momentum

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0046] In order to make the technical problems, technical solutions and beneficial effects to be solved by the embodiments of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0047] Such as figure 1 As shown, the preferred embodiment of the present invention discloses a deep neural network training method based on critical damping momentum, comprising the following steps:

[0048] S1: start a new round of iteration;

[0049] S2: Input a batch of new images, and calculate the trace of the Hessian matrix of the loss function of the neural network;

[0050] Specifically, for each batch of input image data, the trace of the loss function Hessian matrix (ie, the sum of diagonal elements of the Hessian matrix) is calculated...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a training method and system of a deep neural network based on critical damping momentum. The method comprises the following steps: S1, starting a new round of iteration; S2, inputting a batch of new images, and calculating traces of a Hessel matrix of a loss function of the neural network, the Hessel matrix being a matrix formed by second derivatives of the loss function to each parameter of the neural network; S3, substituting the trace of the Hessel matrix into a critical damping solution of a second-order differential equation, and calculating to obtain a momentum coefficient of a neural network weight parameter; S4, updating parameters of the neural network in cooperation with a learning rate attenuation strategy; S5, judging whether all image batches are calculated or not, and if so, executing the step S6; if not, returning to the step S2; and S6, judging whether the iteration turns reach the maximum or not, if so, ending the training, and if not, returning to the step S1. According to the method, a stochastic gradient descent method containing momentum is improved, and an ideal training effect is achieved.

Description

technical field [0001] The invention relates to the technical fields of machine learning and numerical methods, in particular to a training method and system for a deep neural network based on critical damping momentum. Background technique [0002] Deep neural networks have demonstrated excellent performance in fields such as computer vision, natural language processing, and computer graphics. With the increasing complexity of the deep neural network in terms of network layers and topology, the number of parameters of the deep neural network is also increasing day by day. At present, the number of parameters of the latest neural network in the industry has reached 10. 7 ~10 8 Magnitude. Although the huge amount of parameters improves the scope and efficiency of the neural network, it also brings difficulties to the training and optimization of the network. The training of large networks often takes several days or more, and it also puts new demands on the performance of ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063G06N3/08G06K9/62G06F17/16G06F17/13
CPCG06N3/063G06N3/08G06N3/084G06F17/13G06F17/16G06F18/214
Inventor 王好谦蔡宇麟王兴政
Owner SHENZHEN GRADUATE SCHOOL TSINGHUA UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products