Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Intelligence type retrieval dialogue method based on pre-training and attention interaction network

A pre-training and attention technology, applied in the fields of deep learning and natural language processing, can solve the problem of the decline of separation coding accuracy, and achieve the effect of improving the ability of accurate retrieval response, enhancing the ability of representation, and improving the retrieval speed.

Pending Publication Date: 2022-07-29
SOUTH CHINA UNIV OF TECH
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This method is based on the pre-trained language model BERT, which can effectively use the powerful semantic capture ability learned by BERT pre-trained on a large general-purpose corpus, and by continuing to pre-train on the target corpus, use the domain that has learned the semantic representation of the target corpus Adaptive BERT as the encoder; the method also incorporates an interactive network based on cross-attention to alleviate the accuracy drop caused by separate encoding

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Intelligence type retrieval dialogue method based on pre-training and attention interaction network
  • Intelligence type retrieval dialogue method based on pre-training and attention interaction network
  • Intelligence type retrieval dialogue method based on pre-training and attention interaction network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0091] The knowledge-based retrieval dialogue method based on pre-training and attention interaction network in this embodiment is implemented on a role-based dialogue corpus (Persona-Chat). The implementation process includes a domain-adaptive pre-training phase and a fine-tuning training phase.

[0092] figure 1 , figure 2 is an explanatory diagram of the field adaptability pre-training stage during the implementation of the present invention, such as figure 1 , figure 2 As shown, the basic steps of domain-adaptive pre-training implemented on the Persona-Chat corpus are as follows:

[0093] S1. The pre-trained language model of this embodiment selects the basic, case-insensitive BERT model proposed by Google. The BERT model is a neural network including a 12-layer, 768-dimensional, 12 self-attention head, and 110M parameters. Structure; the domain adaptive pre-training hyperparameters are set as follows: the training batch size is 20, the dropout probability is 0.2, th...

Embodiment 2

[0159] The knowledge-based retrieval dialogue method based on pre-training and attention interaction network of the present invention is implemented on a document-based dialogue corpus (CMUDoG). The implementation process includes a domain-adaptive pre-training phase and a fine-tuning training phase.

[0160] figure 1 , figure 2 is an explanatory diagram of the field adaptability pre-training stage in the implementation process of the present invention, such as figure 1 , figure 2 As shown, the basic steps of domain-adaptive pre-training implemented on the CMUDoG corpus are as follows:

[0161] S1. The pre-trained language model of this embodiment selects the basic, case-insensitive BERT model proposed by Google, and the BERT model is a neural network including a 12-layer, 768-dimensional, 12 self-attention head, and 110M parameters Structure; the domain adaptive pre-training hyperparameters are set as follows: the training batch size is 10, the dropout probability is 0....

Embodiment 3

[0217] The knowledge-based retrieval dialogue method based on pre-training and attention interaction network of the present invention is implemented on a document-based dialogue corpus (Persona-Chat). The implementation process includes a domain-adaptive pre-training phase and a fine-tuning training phase.

[0218] figure 1 , figure 2 is an explanatory diagram of the field adaptability pre-training stage in the implementation process of the present invention, such as figure 1 , figure 2 As shown, the basic steps of domain-adaptive pre-training implemented on the Persona-Chat corpus are as follows:

[0219] S1. The pre-trained language model of this embodiment selects the ALBERT model proposed by Google with a faster training speed. The ALBERT model includes a 12-layer, 128-dimensional embedding layer, 128-dimensional hidden layer, 12 self-attention heads, 10M The neural network structure of the parameters; the domain adaptive pre-training hyperparameters are set as follo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a knowledge-based retrieval dialogue method based on pre-training and an attention interaction network, and the method comprises the following steps: training a pre-training language model BERT on a target corpus through employing a domain adaptability pre-training method, and obtaining the domain adaptability BERT; using the field adaptability BERT as an encoder of the attention interaction network, and respectively encoding the dialogue context, the background knowledge and the plurality of candidate response texts to obtain corresponding representations; and finally, respectively inputting the dialogue context, the background knowledge and the representation of the plurality of candidate responses into the attention interaction network for matching, and training the attention interaction network to retrieve the optimal response from the plurality of candidate responses. According to the method, the powerful semantic characterization capability of the pre-training language model is utilized, the semantic characterization capability of the pre-training language model on a specific corpus is improved through two pre-training tasks, and the performance reduction caused by separation coding adopted for improving the retrieval speed is relieved by adopting the attention interaction network.

Description

technical field [0001] The invention relates to the fields of deep learning and natural language processing, in particular to a knowledge-based retrieval dialogue method based on pre-training and attention interaction network. Background technique [0002] Dialogue systems are an important topic in natural language processing, where the goal is to enable computers to understand human dialogue and build end-to-end dialogue devices. There are two kinds of mainstream dialogue systems at present, which are generative dialogue system and retrieval dialogue system. Generative dialogue systems understand dialogue and generate responses through an encoder-decoder structure; retrieval dialogue systems retrieve responses from a corpus. The knowledge-based dialogue response selection task is proposed by the role-based dialogue corpus (Persona-Chat) and the document-based dialogue corpus (CMUDoG), which requires the knowledge-based retrieval dialogue system to select from several candi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/332G06F16/33G06F40/211G06F40/289G06N3/04G06N3/08
CPCG06F16/3329G06F16/3344G06F40/211G06F40/289G06N3/08G06N3/045Y02D10/00
Inventor 苏锦钿陈燕钊
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products