Neural network model training method, audio generation method and device and electronic equipment

A technology of neural network model and training method, which is applied in the field of sound synthesis, can solve problems such as poor sound quality, achieve the effect of improving sound quality and enhancing deep modeling ability

Pending Publication Date: 2020-12-11
TENCENT MUSIC ENTERTAINMENT TECH SHENZHEN CO LTD
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] In the related technology, the method based on parameter synthesis is used to synthesize dry sound, which can achieve the effect of accurate pronunciation time and controllable rhythm under the condition of accurate parameters, but the sound quality is generally poor
It can be seen that in the process of realizing the present invention, the inventors have found that there are at least the following problems in the related art: the sound quality of the synthesized dry sound is relatively poor

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network model training method, audio generation method and device and electronic equipment
  • Neural network model training method, audio generation method and device and electronic equipment
  • Neural network model training method, audio generation method and device and electronic equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0043] The applicant of the present application has found through research that the unvoiced phonetic symbols do not vibrate the vocal cords when they are uttered, and the pronunciation of each unvoiced phonetic symbol is different due to the difference in the subsequent phonetic symbols. For example, for the Chinese characters "春" and "茶", the corresponding pinyin are "chun" and "cha", respectively, which contain the same unvoiced phonetic symbol "ch". When mouth-pronouncing "chun" and "tea", although the unvoiced sounds are the same, the voiced sounds after the unvoiced sounds are different, that is, "un" and "a" are different, resulting in different correspondences to the unvoiced "ch". The mouth shape, that is to say, in different Chinese characters for the same unvoiced sound "ch", the sound is produced in different ways.

[0044] In related technologies, one unvoiced phonetic symbol corresponds to one IPA. In the above example, the IPA corresponding to the unvoiced phone...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a neural network model training method and device, an audio generation method and device, electronic equipment and a computer readable storage medium. The neural network modeltraining method comprises the steps of obtaining a training dry audio, determining all phonetic symbols in the training dry audio, and converting each phonetic symbol into an international phonetic symbol; determining a differentiated international phonetic symbol corresponding to each unvoiced phonetic symbol according to the subsequent phonetic symbol of each unvoiced phonetic symbol in all phonetic symbols and the international phonetic symbol corresponding to each unvoiced phonetic symbol; generating phoneme information corresponding to the training dry audio on the basis of the differentiated international phonetic symbols corresponding to the unvoiced phonetic symbols and the international phonetic symbols corresponding to the phonetic symbols except the unvoiced phonetic symbols; and training a neural network model by using the training dry audio and the corresponding phoneme information so as to establish a mapping relationship between the phoneme information and the dry audio.According to the neural network model training method, the sound quality of the synthesized dry sound is improved.

Description

technical field [0001] The present application relates to the technical field of sound synthesis, and more specifically, to a neural network model training method and device, an audio generation method and device, electronic equipment and a computer-readable storage medium. Background technique [0002] With the development of deep learning technology and audio signal processing technology, artificially synthesized singing voices have gradually become possible. People can use technology to generate dry voices, that is, pure human voices without music. These synthesized dry sounds are accompanied by an accompaniment to obtain a song. [0003] In the related art, a method based on parameter synthesis is used to synthesize dry sound, which can achieve the effect of accurate pronunciation time and controllable rhythm under the condition of accurate parameters, but the sound quality is generally poor. It can be seen that during the process of implementing the present invention, ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G10L13/02G10L13/08G10L25/30G10L25/93
CPCG10L13/02G10L13/08G10L25/30G10L25/93
Inventor 徐东
Owner TENCENT MUSIC ENTERTAINMENT TECH SHENZHEN CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products