Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Voice emotion recognition method based on glottal wave signal feature extraction

A technology of speech emotion recognition and signal characteristics, applied in speech analysis, instruments, etc., can solve the problems of dimension disaster, multi-dimensionality, high redundancy, etc., and achieve less resonance ripples, better recognition effect, and better recognition effect

Inactive Publication Date: 2020-04-28
湖南绚丽新材料科技有限公司
View PDF8 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The glottal wave signal feature extraction is the key to optimizing speech emotion recognition. The traditional method of extracting glottal wave is based on inverse filtering IF (Inverse Filtering). For example, according to the input signal, the free field outside the lips passes through a specially designed breathing air The flow meter records the gas volume velocity in the oral cavity, and obtains the glottal wave through IF. Although it is not easily affected by low-frequency noise, the design of the flow meter has high requirements for the accuracy of the sensor, which is difficult to meet the requirements of ordinary factory production. The extraction method is the residual signal harmonic sum (SRH algorithm), and its accuracy needs to be further broken through. In addition, the traditional speech signal processing technology uses a wide range of emotion-related feature parameters, resulting in too many dimensions and high redundancy. And it is easy to cause the disaster of dimensionality to affect the recognition effect of speech emotion

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Voice emotion recognition method based on glottal wave signal feature extraction
  • Voice emotion recognition method based on glottal wave signal feature extraction

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0017] The technical solutions of the present invention will be further described below in conjunction with embodiments.

[0018] Such as figure 1 As shown, a kind of speech emotion recognition method based on glottal wave signal feature extraction of the embodiment of the present invention comprises the following specific steps:

[0019] Step 1: Speech input and front-end processing. After the speech signal is input, the emotion description model and the CASIA Chinese emotion corpus are expressed in discrete dimensions. After the front-end preliminary processing of the TEO and spectrogram path, a transfer function is given by formula 2.1 The filter is used to realize the pre-emphasis of glottal excitation, and the speech signal is intercepted into data frames with the same length. Generally, the frame length is 10-50ms, and the frame overlap is 5-25ms. Then, based on the unvoiced and voiced sound algorithm of the W-SRH algorithm, the emotional speech signal is discriminated....

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a voice emotion recognition method based on glottal wave signal feature extraction. According to the method, a spectrogram and TEO are mainly adopted as input of CRNN, a low-level descriptor and a high-level statistical function are combined, emotional speech features are extracted, subjected to dimensionality reduction and subjected to an identification algorithm, and finally output is conducted through an HSF channel. The glottal wave signal feature extraction mainly realizes signal acquisition in a complex cepstrum phase decomposition form, a PCA method is adopted toreduce the dimension of a feature vector, and a BP neural network algorithm is adopted to identify and output the feature vector. According to the method, vocal cord vibration characteristics can bebetter reflected, glottis open-phase and closed-phase information is clear, sound source harmonic components and sound channel interference are effectively reduced, and the recognition accuracy is high.

Description

technical field [0001] The invention belongs to the fields of speech recognition, intelligent processing and human-computer interaction, and in particular relates to a speech emotion recognition method based on glottal wave signal feature extraction. Background technique [0002] Speech is widely used as an important information resource transmission and communication medium. The acoustic signal of speech contains a large amount of user information, semantic information and rich emotional information. The development direction of phonetic tasks mainly includes voiceprint recognition and speech recognition. And emotion recognition, speech emotion recognition aims to identify the correct emotional state of the speaker through speech signals. Since speech is not a complete expression of emotional physiological signals, how to efficiently and accurately recognize the emotions expressed by users under the premise of ignoring other sensory results , is a hot topic in phonetic rese...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G10L25/03G10L25/15G10L25/18G10L25/21G10L25/24G10L25/30G10L25/63G10L25/93
CPCG10L25/03G10L25/24G10L25/21G10L25/18G10L25/15G10L25/30G10L25/63G10L25/93
Inventor 易宏博周磊
Owner 湖南绚丽新材料科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products