Self-learning emotion interaction method based on multi-modal recognition

An interaction method and multi-modal technology, applied in the field of human-computer interaction, can solve problems such as insufficient interaction ability

Active Publication Date: 2020-12-15
SOUTH CHINA UNIV OF TECH
View PDF16 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to solve the defect of insufficient interaction ability, and provide a self-learning emotional interaction method based on multi-modal recognition. Comprehensive consideration of fusion features, combined with emotional history state and dialogue memory network, to complete interactive tasks

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Self-learning emotion interaction method based on multi-modal recognition
  • Self-learning emotion interaction method based on multi-modal recognition
  • Self-learning emotion interaction method based on multi-modal recognition

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0068] This embodiment specifically discloses a self-learning emotional interaction method based on multimodal recognition, as shown in the attached figure 1 shown, including the following steps:

[0069] S1. Use the microphone array and the non-contact channel of the camera to collect voice, face and gesture information respectively, as shown in the attached figure 2 As shown in the left half, the technologies used are facial recognition, speech recognition and gesture recognition. Face recognition converts face image signals into face image information, speech recognition extracts voice information from voice signals, and gesture recognition converts gesture image signals into gesture information.

[0070] S2, face image information, voice information and gesture information are processed through a multi-layer convolutional neural network, such as figure 2 As shown in the right part, through emotion analysis technology and under the auxiliary processing of NLP, the speec...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a self-learning emotion interaction method based on multi-modal recognition. The method comprises the following steps: respectively collecting voice, face and gesture signals by a non-contact channel; performing feature extraction on the signals to obtain preliminary features of the signals; inputting the features into a bidirectional LSTM layer to obtain single-mode private information and multi-mode interaction information, and obtaining fusion features according to the information; predicting the emotion of a user based on a classification learning algorithm in combination with the multi-modal fusion features and a historical emotional state curve, and selecting an interaction mode; in the interaction mode, giving an interaction response according to the dialoguememory network; and finally, feeding back and optimizing the emotional state curve and the dialogue memory network according to the interaction effect. According to the method, an operator is allowedto input information through multiple channels of a non-contact man-machine interaction interface, multi-modal fusion features are comprehensively considered, and an interaction task is completed incombination with the historical emotional state and the dialogue memory network.

Description

technical field [0001] The invention relates to the technical field of human-computer interaction, in particular to a self-learning emotional interaction method based on multi-modal recognition. Background technique [0002] Intelligent human-computer interaction is an important direction for the development of artificial intelligence. With the development of mobile Internet, higher requirements are put forward for the human-like and natural nature of human-computer interaction. [0003] The current interaction technology is relatively simple, mostly pure text or voice interaction. Some so-called multi-modal interaction methods are only simple addition operations on multi-modal features, which are separate processing of multiple single-modal information. The interaction information between multimodal information causes ambiguity between multiple modalities, and it is impossible to achieve a complete and unambiguous interaction task. At the same time, most of the interaction...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F3/01G06F3/16G06K9/00G06K9/62G06N3/04
CPCG06F3/011G06F3/017G06F3/167G06V40/113G06V40/168G06N3/044G06N3/045G06F18/253
Inventor 潘粤成刘卓潘文豪邓晓燕蔡典仑
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products