Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Bimodal fusion emotion recognition method based on video image and voice

An emotion recognition and video image technology, applied in the field of emotion recognition, can solve the problems of difficult acquisition of physiological parameters, difficulty in improving recognition accuracy, and manual feature extraction, so as to optimize nonlinear processing capabilities, realize real-time emotion recognition, and achieve accurate recognition. The effect of rate increase

Pending Publication Date: 2021-09-03
NANJING UNIV OF TECH
View PDF7 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Due to the difficulty of collecting physiological parameters, this analysis method is rarely used; the accuracy of body movement recognition is low, and it usually appears as an auxiliary recognition method; the difficulty of collecting voice and facial expressions is not high, but the recognition effect is good, which is practical application The most extensive emotion recognition method
[0003] The currently used emotion recognition methods mainly have the following shortcomings: many single-modal recognition methods are used, and the accuracy of single-modal recognition is difficult to continue to improve; manual feature extraction is required, and real-time processing cannot be achieved; fusion of different modalities mostly uses feature fusion technology , resulting in higher feature dimensions and unable to achieve real-time processing
These shortcomings make it difficult to improve the recognition accuracy to a higher level, and it is also difficult to realize real-time emotion recognition, so it is necessary to improve

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Bimodal fusion emotion recognition method based on video image and voice
  • Bimodal fusion emotion recognition method based on video image and voice
  • Bimodal fusion emotion recognition method based on video image and voice

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0014] The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application.

[0015] The invention provides a dual-mode fusion emotion recognition method based on video image and voice, the method is as follows: the image training data set is input into the improved convolutional neural network model for training to obtain the video image modal model; the voice training data set is Input the improved long-term and short-term memory neural network model for training to obtain the voice mode model; collect real-time video images and language information from the camera and microphone, and send them to the emotion recognition unit; the emotion recognition unit includes video image mode and voice mode , the recognition results of the two modalities are respectively obtained, and the recognition process is carried out on the trained neural network model; the recognition results are...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a bimodal fusion emotion recognition method based on video images and voice. The method is composed of a camera, a microphone and an emotion recognition unit, wherein the emotion recognition unit is composed of a video image mode and a voice mode. The training process of the bimodal model is as follows: inputting an image training data set into a convolutional neural network model for training to obtain a video image modal model; and inputting the voice training data set into the long-short-term memory neural network model for training to obtain a voice modal model. The camera collects video images and sends the video images to the emotion recognition unit, and the facial expression features are analyzed to obtain recognition results; the microphone collects voice data and sends the voice data to the emotion recognition unit, and analyzes voice emotion features to obtain a recognition result; and fusing the identification results of the two modals at a decision-making layer according to a weight criterion and outputting the fused identification results. The recognition method can improve the accuracy of emotion recognition and realize real-time detection.

Description

technical field [0001] The invention relates to the field of emotion recognition, in particular to a dual-modal fusion emotion recognition method based on video image and voice. Background technique [0002] With the rapid development of artificial intelligence technology, people hope that there will be a more vivid interaction between AI and users, so as to bring users a better user experience. In terms of engineering application value, emotion recognition is a research topic involving machine vision, medicine, psychology and many other fields. The research can not only promote the progress of other interdisciplinary subjects, but also bring huge commercial value and practical significance to the society. . According to different analysis information, emotion recognition technology can be divided into two categories at present; one is based on physiological signals, such as EEG, ECG, etc.; the other is based on the analysis of emotional behavior, such as facial expressions...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04G06N3/08G10L25/30G10L25/63
CPCG06N3/084G10L25/30G10L25/63G06N3/045G06F18/253
Inventor 李为相王传昱程明
Owner NANJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products