Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

RGB-D saliency detection method based on dynamic filtering decoupling convolutional network

A RGB-D, dynamic filtering technology, applied in the field of computer vision

Pending Publication Date: 2021-09-10
DALIAN UNIV OF TECH
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, reusing fusion modules inevitably incurs additional computation and storage consumption
This further extends the challenge of cross-modal feature fusion

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • RGB-D saliency detection method based on dynamic filtering decoupling convolutional network
  • RGB-D saliency detection method based on dynamic filtering decoupling convolutional network
  • RGB-D saliency detection method based on dynamic filtering decoupling convolutional network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0055] The object of the present invention is to provide a RGB-D saliency detection method based on a dynamic filter decoupling convolutional network to effectively fuse RGB and depth two modal information to achieve high-quality saliency detection results in complex scenes . The first challenge faced by the object of the present invention is to design a module that can dynamically adapt to the specific presence in each modality, and the second challenge is to dynamically build complementary interactions for inter-modal fusion.

[0056] The core idea of ​​the present invention is to design a decoupled dynamic filtering saliency detection network to dynamically promote feature interaction by decoupling the dynamic convolution to the spatial dimension and the channel dimension to deal with both intra-modal and inter-modal issues .

[0057] In order to enable those skilled in the art to better understand the solutions of the present invention, the following will clearly and comp...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an RGB-D saliency detection method based on a dynamic filtering decoupling convolutional network. The method comprises the following steps: acquiring an RGB image tensor and a depth image tensor; respectively inputting the RGB image tensor and the depth image tensor into an encoder network to extract a single-mode feature group, and generating an RGB feature group and a depth feature group according to the characteristics of the encoder network and hierarchical division; respectively putting the features of the RGB feature group and the features of the depth feature group into a respective modal-specific overall guide dynamic enhancement module (MGDEM), and performing single-modal specific feature enhancement; inputting the enhanced RGB features and depth features into a scene perception cross-modal dynamic fusion module (SCDFM), and performing feature fusion between modals, wherein the MGDEM and the SCDFM are both based on a decoupled dynamic filtering convolution structure; and inputting the fused features into a decoder to obtain a predicted significance result. According to the method, accurate saliency prediction is realized.

Description

technical field [0001] The invention relates to the technical field of computer vision, in particular to an RGB-D saliency detection method based on a dynamic filter decoupling convolutional network. Background technique [0002] The goal of saliency detection is to find the object or region that most attracts human visual attention in an image. This task plays a preprocessing role in different vision tasks, allowing the computer to first detect areas of interest to humans, and then perform subsequent processing on the detected areas. For example, when performing target recognition, it is not necessary to traverse the entire image with a sliding window, but to perform saliency detection first, and then only recognize the target in the saliency area. Accurate and reliable saliency detection can save computation while increasing accuracy, benefiting many image tracking and recognition processing tasks in visual graphics. [0003] Many current methods perform saliency detecti...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06T7/00G06T7/33G06T7/50G06K9/62G06N3/04G06N3/08
CPCG06T7/0002G06T7/33G06T7/50G06N3/08G06T2207/10024G06T2207/20081G06T2207/20084G06T2207/20221G06N3/045G06F18/253G06F18/214
Inventor 张淼朴永日姚舜禹
Owner DALIAN UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products