Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-round dialogue device and method

A feature vector, the above technology, applied in semantic analysis, instruments, electronic digital data processing and other directions, can solve problems such as slow speed, lack of dialogue data, lack of contextual dialogue information for semantic features, etc., to achieve accurate prediction, simple network structure, The effect of improving natural language understanding ability

Pending Publication Date: 2021-11-19
广州天宸健康科技有限公司
View PDF8 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] The general pre-training language model, such as the bert model, is a multi-layer bidirectional transformer network structure, which performs self-supervised learning on the basis of massive corpus. The feature representation obtained through the bert model greatly improves the accuracy of natural language processing tasks. , but each layer of the bert model is self-supervised, resulting in an overall complexity of O(n2), which requires a lot of machine resources
[0003] In a multi-round and context-dependent chat system, the bert model is not ideal. In addition to the large amount of calculation, slow speed, and high training cost, its biggest defect is that the bert model is a model based on general corpus training, because the general corpus The learned semantic features lack strongly relevant contextual dialogue information, and most of the corpus is based on documents, lacking dialogue data. Therefore, using the bert model for multiple rounds of dialogue cannot improve the robot's natural language understanding ability and accuracy of intention judgment.
Especially in the colloquial, specific scene or professional industry knowledge field, in the multi-round chat system with semantic correlation between the upper and lower sentences, the expressive power is limited and the accuracy rate is not high

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-round dialogue device and method
  • Multi-round dialogue device and method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment approach

[0054] As a preferred embodiment, its implementation is as follows:

[0055] First the words of each sentence of the above dialogue text data, question data and answer data are segmented, then the position or address of each word is represented by ID, and the ID information of each word is used in N dimensions (such as 512 dimensions) random vector representation, thereby constructing the sentence vector set of the above dialogue text data, question data and answer data, and then can use this sentence vector set to input into the feature extraction module for feature extraction or extraction.

[0056]In this embodiment, the position or address of each word is represented by an ID, so that in the multi-round dialogue device of the present application, it is used to select masking for predictive training during pre-training. In the specific implementation, a certain probability algorithm can be designed, each ID has a certain probability or is randomly selected to be replaced by...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-round dialogue device and method.The multi-round dialogue device comprises a data processing module, a characterization module, a feature extraction module, a question and answer feature similarity module and an objective function module. The data processing module is used for analyzing multi-round dialogue data of historical chats to obtain input data; the representation module is used for mapping input data to obtain a sentence vector set; the feature extraction module is used for analyzing the sentence vector set; the question and answer feature similarity module is used for processing the sentence vector set to obtain a score matrix; the objective function module is used for setting an objective function suitable for the multi-round dialogue device according to the scoring matrix. Under the condition that the sample size is not large, the multi-round dialogue device can learn good context features, so that the questions of the user can be predicted more accurately and answers can be provided, the network structure is simple, and a lightweight memory and an energy-saving model are realized.

Description

technical field [0001] The invention relates to the field of natural language processing, in particular to a multi-round dialogue device and method. Background technique [0002] The general pre-training language model, such as the bert model, is a multi-layer bidirectional transformer network structure, which performs self-supervised learning on the basis of massive corpus. The feature representation obtained through the bert model greatly improves the accuracy of natural language processing tasks. , but each layer of the bert model is self-supervised, resulting in an overall complexity of O(n2), which requires a lot of machine resources. [0003] In a multi-round and context-dependent chat system, the bert model is not ideal. In addition to the large amount of calculation, slow speed, and high training cost, its biggest defect is that the bert model is a model based on general corpus training, because the general corpus The learned semantic features lack strongly relevant...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/332G06F16/33G06F40/194G06F40/30
CPCG06F16/3329G06F16/3344G06F16/3343G06F16/3346G06F40/194G06F40/30Y02D10/00
Inventor 曾祥云朱姬渊
Owner 广州天宸健康科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products