Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Machine translation model optimization method based on Transform model

A technology of machine translation and optimization methods, applied in the field of evolutionary computing, can solve problems such as difficult to independently design models

Active Publication Date: 2021-06-29
SICHUAN UNIV
View PDF4 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Although the Transformer model has achieved good results in machine translation, there are still several problems: 1. The arrangement mode of the MHA layer and the FFN layer in the Transformer model with different network layers is fixed, and existing studies have shown that Transformer's different layer arrangement modes have better performance than the base Transformer model on other natural language processing tasks
3. The number of layers and hyperparameters of the Transformer model are set by experts combined with domain knowledge. If non-professionals want to use the Transformer model to solve machine translation tasks, it is difficult to independently design a model that meets expectations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Machine translation model optimization method based on Transform model
  • Machine translation model optimization method based on Transform model
  • Machine translation model optimization method based on Transform model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0061] The specific embodiments of the present invention are described below so that those skilled in the art can understand the present invention, but it should be clear that the present invention is not limited to the scope of the specific embodiments. For those of ordinary skill in the art, as long as various changes Within the spirit and scope of the present invention defined and determined by the appended claims, these changes are obvious, and all inventions and creations using the concept of the present invention are included in the protection list.

[0062] A machine translation model optimization method based on the Transformer model, such as figure 1 shown, including the following steps:

[0063] S1. Initialize the population of Transformer models with multiple different structures and parameters;

[0064] Specifically, genetic coding is to express the model as a searchable individual, laying the foundation for subsequent evolutionary search. In order to allow the T...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a machine translation model optimization method based on a Transform model, and the method comprises the steps: enabling Transform individuals to have different structures and parameters through designing variable length codes and candidate blocks, and providing a plurality of candidate models for the learning of word vectors; then designing a crossover and mutation strategy to enable Transform individuals to perform information exchange, and enabling excellent structures or parameters for processing word vectors to be inherited to the next generation; then designing an environment selection strategy to generate next generation Transform individuals, eliminating models with relatively poor word vector learning effects, and reserving models with relatively excellent word vector learning effects; and then finding a transformer model with an optimal word vector learning effect through continuous iterative evolution search for finally solving a machine translation task so that the transformer model can better learn word vector expression in the machine translation task, and the precision of the machine translation task is improved.

Description

technical field [0001] The invention relates to the field of evolutionary computing, in particular to a method for optimizing a machine translation model based on a Transformer model. Background technique [0002] Transformer is a sequence-to-sequence proposed by Google in 2017 to solve machine translation tasks. Before Transformer was proposed, machine translation models can be divided into two categories: models based on feedback neural networks or convolution-based sequence regression Model. Most of the models based on the feedback neural network are composed of RNN or LSTM structure. The input of each layer in the model depends on the output state of the previous layer. The process requires a lot of time and computing resources; the convolution-based network model is composed of a multi-layer convolutional neural network, and the number of convolution operations in this model will increase rapidly when calculating the relationship of long-distance information, such as i...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F40/58G06F40/284G06F40/253G06F40/30G06N3/00
CPCG06F40/58G06F40/284G06F40/253G06F40/30G06N3/006
Inventor 孙亚楠冯犇吴杰李思毅
Owner SICHUAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products