Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-modal emotion analysis method based on multi-dimensional attention fusion network

A technology that integrates network and sentiment analysis, applied in the field of multi-modal emotional computing, can solve problems such as unfavorable production and living environment, model over-fitting, and failure to consider inter-modal correlation information

Active Publication Date: 2020-09-18
HUAZHONG UNIV OF SCI & TECH
View PDF7 Cites 26 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This method is only a simple result integration of the three modalities, and does not take into account the correlation information between the modalities, and it is easy to cause model overfitting due to information redundancy.
Another approach is based on modal labeling alignment, that is, when performing data labeling, the three modalities are forced to align in the time dimension based on text or phonemes, thus ensuring the corresponding relationship of the three modalities in time. Then use the recurrent neural network, convolutional neural network, attention mechanism, and Seq2Seq framework for modal fusion. This method is expensive to label and is not conducive to the actual production and living environment.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal emotion analysis method based on multi-dimensional attention fusion network
  • Multi-modal emotion analysis method based on multi-dimensional attention fusion network
  • Multi-modal emotion analysis method based on multi-dimensional attention fusion network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0076] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0077] The present invention provides a multi-modal emotion analysis method based on a multi-dimensional attention fusion network, the specific process is as follows figure 1 ,in addition figure 2 is a schematic structural diagram of a multi-dimensional attention fusion network in an embodiment of the present invention, image 3 is a schematic structural diagram of the cross-modal fusion module in the embodiment of the present invention. The realization steps of the inventive method are as follows:

[0078] 1. Process the multi-modal emotion database and perform feature dimension alignment.

...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-modal emotion analysis method based on a multi-dimensional attention fusion network. The multi-modal emotion analysis method comprises the steps: extracting voice preprocessing features, video preprocessing features and text preprocessing features from sample data containing multiple modals such as voice, video and text; then, constructing the multi-dimensional attention fusion network for each mode; extracting first-level autocorrelation features and second-level autocorrelation features by using an autocorrelation feature extraction module in the network, thencombining the autocorrelation information of the three modes, and obtaining cross-modal fusion features of the three modes by using a cross-modal fusion module in the network; combining the secondaryautocorrelation features and the cross-modal fusion features to obtain modal multi-dimensional features; and finally, splicing the modal multi-dimensional features, determining emotion scores, and performing emotion analysis. According to the method, feature fusion can be effectively carried out in a non-aligned multi-modal data scene, and multi-modal associated information is fully utilized to carry out emotion analysis.

Description

technical field [0001] The invention belongs to the field of multi-modal emotion computing, and more specifically relates to a multi-modal emotion analysis method based on a multi-dimensional attention fusion network. Background technique [0002] Sentiment analysis has many applications in everyday life. With the development of big data and multimedia technology, the use of multi-modal sentiment analysis technology to analyze the different modes of data voice, video, and text is more conducive to mining the shallow meaning behind the data. For example, in the return visit survey, through the comprehensive analysis of the user's voice, face and speech content, we can know the user's satisfaction with the service or product. [0003] At present, the difficulty of multimodal sentiment analysis lies in how to effectively integrate multimodal information, because the acquisition methods of voice, video and text features are completely different. When describing the same conten...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06F40/30G10L25/63G06N3/04
CPCG06F40/30G10L25/63G06V40/174G06N3/045G06F18/253
Inventor 冯镔付彦喆王耀平江子文杭浩然李瑞达刘文予
Owner HUAZHONG UNIV OF SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products