Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network model compression method, corpus translation method and devices thereof

A technology of neural network model and compression method, which is applied in the field of corpus translation method and its device, and neural network model compression of machine translation, which can solve problems such as difficulty in obtaining neural network student models, and achieve the effect of improving quality

Active Publication Date: 2020-02-21
BEIJING XIAOMI INTELLIGENT TECH CO LTD
View PDF8 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] However, in the process of knowledge distillation training, a teacher model is used to guide a student model, which is a single-stage knowledge distillation method. Since the prediction accuracy of the teacher model is generally much better than that of the student model, it is difficult to integrate the knowledge of the teacher model through single-stage knowledge distillation. Efficient transfer to the student model, so it is difficult to obtain a neural network student model with better prediction accuracy for mobile deployment

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network model compression method, corpus translation method and devices thereof
  • Neural network model compression method, corpus translation method and devices thereof
  • Neural network model compression method, corpus translation method and devices thereof

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0042] The exemplary embodiments will be described in detail here, and examples thereof are shown in the accompanying drawings. When the following description refers to the accompanying drawings, unless otherwise indicated, the same numbers in different drawings represent the same or similar elements. The implementation manners described in the following exemplary embodiments do not represent all implementation manners consistent with the present invention. Rather, they are merely examples of devices and methods consistent with some aspects of the present invention as detailed in the appended claims.

[0043] In order to achieve the training effect of the student model, an embodiment of the present disclosure provides a neural network model compression method 10, see figure 1 , A neural network model compression method 10, including steps S11-S15, each step is described in detail below:

[0044] Step S11: Obtain a training sample set including a plurality of training sample pairs,...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a neural network model compression method, a corpus translation method and devices thereof. The neural network model compression method comprises the steps of acquiring a training sample set comprising a plurality of training sample pairs, wherein each training sample pair comprises source data and target data corresponding to the source data; training an original teachermodel by taking the source data as input and taking the target data as verification; training an intermediate teacher model based on the training sample set and the original teacher model, and forminga teacher model set by one or more intermediate teacher models; training a plurality of candidate student models based on the training sample set, the original teacher model and the teacher model set, and forming a student model set by the plurality of candidate student models; and according to the training results of the plurality of candidate student models, selecting one candidate student model as a target student model. By introducing a plurality of teacher models, multistage guidance of student model training is realized, so that the quality of the student model is improved.

Description

Technical field [0001] The present disclosure relates to the field of machine translation, and in particular to neural network model compression methods, corpus translation methods and devices for machine translation, as well as electronic devices and computer-readable storage media. Background technique [0002] With the development of artificial intelligence technology, deep learning based on neural networks has achieved good performance on multiple machine learning tasks such as image classification, machine translation, and speech recognition. [0003] At present, in order to obtain the best network model output results, a deep neural network model containing multiple network layers and ultra-large-scale parameters is usually used. Although this complex network structure significantly improves the output results of the model, it also makes it difficult to Deployment on mobile devices with small storage space, on the other hand, also leads to excessive inference delays on low-po...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F40/58G06N3/08
CPCG06N3/08G09B23/10G06N3/045G06N3/04G06F40/58G06F40/30G06F18/214
Inventor 李响孙于惠姜佳良崔建伟
Owner BEIJING XIAOMI INTELLIGENT TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products