Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

BERT embedded speech translation model training method, BERT embedded speech translation model training system, speech translation method and speech translation equipment

A technology of speech translation and model training, applied in the field of speech translation, can solve the problems affecting the efficiency and quality of speech translation and the inability to use neural networks, etc., so as to improve the quality and application effect of speech translation, good application prospects, improve training efficiency and recognition performance effect

Active Publication Date: 2021-04-20
PLA STRATEGIC SUPPORT FORCE INFORMATION ENG UNIV PLA SSF IEU +1
View PDF5 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

For the neural network, even a network with relatively poor effect may produce good results for a certain input, it is obviously impossible to use the uncertainty of the neural network to guide the training with a single teacher model, which will affect the efficiency and quality of speech translation.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • BERT embedded speech translation model training method, BERT embedded speech translation model training system, speech translation method and speech translation equipment
  • BERT embedded speech translation model training method, BERT embedded speech translation model training system, speech translation method and speech translation equipment
  • BERT embedded speech translation model training method, BERT embedded speech translation model training system, speech translation method and speech translation equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] In order to make the purpose, technical solution and advantages of the present invention more clear and understandable, the present invention will be further described in detail below in conjunction with the accompanying drawings and technical solutions.

[0030] The embodiment of the present invention provides a BERT embedded speech translation model training method based on entropy weighted knowledge distillation, which includes the following content: collecting model training data, the training data includes but not limited to: the source language of the speech translation data is paired with the target language Text; use the source language to pre-train the BERT model, use the pre-trained BERT model as the encoding layer of the machine translation model, and use the paired source language and target language texts to train the machine translation model, and set the machine translation model to decode The number of layers is used to obtain multiple machine translation...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the technical field of speech translation, and relates to a BERT embedded speech translation model training method and system and a speech translation method and device.The training method comprises the following steps: collecting model training data; pre-training a BERT model by utilizing a source language in the training data, taking the pre-trained BERT model as a machine translation model coding layer, training a machine translation model by utilizing paired source language and target language texts, and obtaining a plurality of machine translation models by setting the number of decoding layers in the machine translation model; training a speech recognition model by using the speech translation data of the source language pairs; and taking the trained speech recognition model coding layer as a speech translation model coding layer initialization parameter, weighting the outputs of the plurality of machine translation models in an entropy weighting mode to train a speech translation model, and completing speech translation model training in combination with a model loss function. The recognition performance of the speech translation model is improved, and then the speech translation efficiency and quality are improved.

Description

technical field [0001] The invention belongs to the technical field of speech translation, and in particular relates to a BERT embedded speech translation model training method and system based on entropy weighted knowledge distillation, a speech translation method and equipment. Background technique [0002] Speech translation is an end-to-end model that converts speech in one language into text in another language, that is, skips the step of converting the source language into text in the traditional model, and directly converts the speech in the source language into the target Language text is a hot research topic at present. In terms of model selection, the transformer model proposed by Google has become the main choice due to the effectiveness of its self-attention mechanism modeling, the efficiency of parallel processing, and the simplicity of the model structure, which has significantly improved the recognition rate. But one of the main problems is the amount of data...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/58G10L15/06G10L15/22
Inventor 屈丹张昊杨绪魁张连海陈琦李真李静涛
Owner PLA STRATEGIC SUPPORT FORCE INFORMATION ENG UNIV PLA SSF IEU
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products