Supercharge Your Innovation With Domain-Expert AI Agents!

Target tracking method for RGB-D (RGB-Depth) data cross-modal feature learning based on sparse deep denoising autoencoder

A mode feature and self-encoder technology, applied in the field of target tracking, can solve the problems of ignoring the correlation between RGB mode and Depth mode, and achieve high accuracy and strong robustness

Active Publication Date: 2016-11-16
SHANGHAI GUAN AN INFORMATION TECH
View PDF2 Cites 23 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Depth information can improve the performance of video object tracking. How to better integrate RGB information and depth information to improve the performance of video object tracking is the focus of research in the field of video object tracking. The following technical problems generally exist in previous related technologies: 1. Both the HOG feature and the Harr feature are artificially designed features, which have certain limitations; 2. These related technologies operate on the artificially designed features in RGB mode and Depth mode (depth mode), and then use weighted and other methods simply fuse the features extracted in the two modes, and the fusion method ignores the complex correlation between the RGB mode and the Depth mode

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Target tracking method for RGB-D (RGB-Depth) data cross-modal feature learning based on sparse deep denoising autoencoder
  • Target tracking method for RGB-D (RGB-Depth) data cross-modal feature learning based on sparse deep denoising autoencoder
  • Target tracking method for RGB-D (RGB-Depth) data cross-modal feature learning based on sparse deep denoising autoencoder

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0037] like image 3 , Figure 4 , Figure 5 , Figure 6 and Figure 13 A target tracking method for cross-modal feature learning of RGB-D data based on sparse depth denoising self-encoder is shown, including the following steps:

[0038] Step 1: Construct an RGB-D cross-modal feature deep learning network with a sparse-limited denoising autoencoder;

[0039] Step 2: collect the RGB-D video unlabeled sample set; the RGB-D video unlabeled sample set includes unlabeled RGB image samples and unlabeled Depth image samples;

[0040] Step 3: using an unsupervised learning method and the RGB-D video unlabeled sample set to train the RGB-D cross-mode feature deep learning network;

[0041] Step 4: For the RGB-D video sequence that contains the tracking target, select the positive and negative samples of the tracking target in the initial m frames as the initial template in the target sample library; the initial template includes the positive and negative RGB images of the trackin...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a target tracking method for RGB-D (RGB-Depth) data cross-modal feature learning based on a sparse deep denoising autoencoder. The method comprises the steps of establishing a RGB-D cross-modal feature deep learning network with a sparse limitation denoising autoencoder; training the RGB-D cross-mode feature deep learning network by employing an unsupervised learning method; inputting RGB image positive and negative samples and Depth image positive and negative samples into the trained RGB-D cross-modal feature deep learning network, thereby obtaining cross-modal features; sending the cross-modal features to a logic regression classifier, training the logic regression classifier by employing a supervised learning algorithm; generating a particle set according to a state transition model; sending the cross-modal features of each particle to the trained logic regression classifier, thereby obtaining confidence scores as observation likelihood modals; and obtaining a posteriori probability of a tth frame in a RGB-D video sequence by employing a particle filtering method, thereby obtaining target tracking results X<^><t>. The method is relatively high in accuracy.

Description

technical field [0001] The invention relates to a target tracking method, in particular to a target tracking method based on RGB-D data cross-mode feature learning of sparse depth denoising self-encoder. Background technique [0002] Video object tracking technology is one of the key technologies in the field of computer vision, and it has been widely used in intelligent video surveillance, robot navigation, video measurement and other fields. The video target tracking methods in the prior art mainly realize target positioning in a two-dimensional image sequence. Due to the lack of three-dimensional information, tracking failures will occur when the target is occluded, rotated, or changes in attitude. With the introduction of depth sensors such as Microsoft Kinect, Asus Xtion, and PrimeSense, it has become a reality to obtain color images and depth images (RGB-D, RGB-Depth data) at the same time. Depth information can improve the performance of video object tracking. How to...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T7/20G06K9/62
CPCG06T7/20G06T2207/10024G06T2207/10028G06T2207/10016G06T2207/20081G06T2207/20084G06F18/24133
Inventor 姜明新
Owner SHANGHAI GUAN AN INFORMATION TECH
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More