Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A real-time voice-driven face animation method and system

A voice and animation technology, applied in the field of real-time voice-driven face animation, can solve the problem of voice animation method relying on the speaker, speaking style, unable to animate retargeting to any facial equipment, etc., to achieve easy editing and high fidelity. Effect

Active Publication Date: 2021-03-19
北京中科深智科技有限公司
View PDF5 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In order to solve the above-mentioned technical problems, the object of the present invention is to provide a method and system for real-time voice-driven facial animation that has nothing to do with the speaker and can be retargeted to any animation equipment, so as to solve the problem that existing voice animation methods rely on Issues with specific speakers, speaking styles, and not being able to retarget generated animations to any facial rig

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A real-time voice-driven face animation method and system
  • A real-time voice-driven face animation method and system
  • A real-time voice-driven face animation method and system

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0060] In order to enable those skilled in the art to better understand the solutions of the present invention, the following will clearly and completely describe the technical solutions in the embodiments of the present invention in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are only It is an embodiment of a part of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts shall fall within the protection scope of the present invention.

[0061] It should be noted that the terms "first" and "second" in the description and claims of the present invention and the above drawings are used to distinguish similar objects, but not necessarily used to describe a specific sequence or sequence. It is to be understood that the data so used are interchangeable under appropriate ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a method and system for real-time voice-driven facial animation, the method comprising acquiring a neutral voice audio-visual data set from a first speaker; using an active appearance model to track and parameterize human face video data; Convert data to phoneme label sequences; train sliding window based deep convolutional neural network model; retarget reference face model to target character model; input retarget phoneme label sequence from second speaker to target role model Deep Convolutional Neural Network Models for Prediction. The system provided by the invention includes: an acquisition module, a face conversion module, a phoneme conversion module, a training module, a retargeting module and a target prediction module. The method and system provided by the present invention solve the problem that the existing voice animation method depends on a specific speaker and speaking style, and cannot retarget the generated animation to any facial equipment.

Description

technical field [0001] The invention relates to the fields of virtual reality and animation, in particular to a method and system for real-time voice-driven facial animation. Background technique [0002] Voice animation is an important and time-consuming aspect of generating photorealistic animation. Broadly speaking, voice animation refers to moving the facial features of a graphic (or robotic) model so that lip movements are synchronized with speech and give the impression of speech generation. As humans, we're all experts on faces, and poor voice animation can be distracting, unpleasant, and confusing. For example, audio-visual language mismatches can sometimes change what viewers think they hear, and high-fidelity voice animation is critical to effective character animation. [0003] However, existing machine learning-based speech animation methods are usually evaluated based on test samples distributed in the same distribution as the training set, and the results are...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06T13/20G06T13/40G06N3/04G06N3/08G10L13/08G10L15/02G10L15/25G10L15/26G10L25/30G10L25/45
CPCG06T13/205G06T13/40G06N3/08G10L15/02G10L15/25G10L25/30G10L25/45G10L13/08G06T2207/10016G06T2207/20081G06T2207/20084G06T2207/30201G10L2015/025G10L15/26G06N3/045
Inventor 不公告发明人
Owner 北京中科深智科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products