Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Human body posture estimation method based on joint relation

A technology of human body posture and joints, applied in computing, computer parts, instruments, etc., can solve problems such as inability to train, a large number of calculations, increase the difficulty of prediction, etc., and achieve the effect of good recognition effect, high computing efficiency, and accurate positioning.

Active Publication Date: 2021-09-24
TONGJI UNIV
View PDF4 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This approach cannot be trained end-to-end, and the graph convolutional network stage is computationally intensive
[0005] The human body is a non-rigid body, and the rotation of each joint has a great degree of freedom. The free rotation of multiple joints can be superimposed on each other. The joints at the far end of the limbs, such as wrist joints and ankle joints, also have multiple degrees of freedom. The location changes, which increases the difficulty of prediction, but the existing methods do not pay attention to the difference in the detection difficulty of different types of joint points brought about by the special structure of the human body; at the same time, in the multi-person pictures that are closer to the real scene, due to Interaction with people, the diversity of positional relationships and the complexity of life scenes, the position of key points of the human body and body information are all facing serious occlusion problems
The occluded human joints are also part of the human pose, and the existing methods do not take targeted optimization measures for the invisible joints that are more difficult to detect correctly

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Human body posture estimation method based on joint relation
  • Human body posture estimation method based on joint relation
  • Human body posture estimation method based on joint relation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0063] Such as figure 2 As shown, a kind of human body posture estimation method based on joint relationship provided by the present invention mainly includes the following four steps:

[0064] 1) Construct the joint relationship module to generate supplementary features for auxiliary positioning of difficult joint points, including two sub-modules: the channel-based feature relationship module and the adjacent joint space relationship module;

[0065] 2) Construct a human body pose estimation model based on joint relationships based on the general deep convolutional neural network model;

[0066] 3) Use the marked human body posture data to train the constructed human body posture estimation model based on the joint relationship, and obtain a network model that can better locate limb joints and occluded joints;

[0067] 4) For the input image to be processed, use the human body pose estimation network with joint relationship module trained in step 3) to perform the human bo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a human body posture estimation method based on a joint relation, and the method comprises the following steps: S1, constructing a joint relation module which comprises two sub-modules: a channel-based feature relation module and an adjacent joint space relation module; s2, constructing a human body posture estimation model based on a joint relation; s3, training a human body posture estimation model based on a joint relation by utilizing the marked human body posture data; and S4, performing a human body posture estimation task based on a single image by using the trained human body posture estimation model added with the joint relation module to obtain a predicted human body posture. Compared with the prior art, the invention effectively overcomes the problem that the positions of four-limb joints with high degree of freedom, such as wrist joints, ankle joints and shielded invisible joints, are difficult to detect in the image, and is high in human body posture estimation accuracy.

Description

technical field [0001] The invention relates to the field of human body pose estimation, in particular to a method for estimating human body pose based on joint relationships. Background technique [0002] Human pose estimation is a traditional task in the field of computer vision. Human pose estimation includes the detection of human key points and the generation of human poses. The "key points" in human key point detection refer to important joints such as the top of the human body, shoulders, elbow joints, wrist joints, and ankle joints. The generated human body posture is the complete human skeleton information. With the innovation of computer vision technology, human pose estimation has also gone through a process from manually extracting features to using deep convolutional neural networks as a tool. In recent years, the development of basic deep convolutional neural network structure and performance has also greatly improved the level of extracting human joint featu...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/62G06N3/04
CPCG06N3/045G06F18/214
Inventor 梁爽储港谢驰王颉文
Owner TONGJI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products