Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Indoor scene semantic annotation method

A technology for semantic labeling and indoor scenes, applied in the fields of multimedia technology and computer graphics, can solve the problems of not increasing the receptive field of feature learning, and not improving the feature expression of feature learning.

Active Publication Date: 2019-08-09
BEIJING UNIV OF TECH
View PDF14 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, these methods decouple context modeling from convolutional feature learning, which may pose the problem of finding suboptimal solutions due to small differences in feature representations.
Another type of method uses a cascaded recurrent neural network with a gate structure, such as a long short-term memory (LSTM) network, to strengthen context modeling, but the method of building a context model based on a recurrent neural network only fuses context information at a specific level of the network. , did not increase the receptive field in the feature learning process, nor did it improve the feature expression in the feature learning process

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Indoor scene semantic annotation method
  • Indoor scene semantic annotation method
  • Indoor scene semantic annotation method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0037] like figure 1 , 2 As shown, this indoor scene semantic annotation method includes the following steps:

[0038] (1) Input color picture and depth picture;

[0039] (2) Entering the neural network, the color picture and the depth picture pass through conv1 and conv2_x respectively;

[0040] (3) Enter the first attention mechanism module ARF_1, and obtain the feature map through the calculation of ARF_1;

[0041] (4) Enter conv3_x for convolution calculation;

[0042] (5) Enter the second attention mechanism module ARF_2, and obtain the feature map through the calculation of ARF_2;

[0043] (6) Enter conv4_x to perform hole convolution calculation;

[0044] (7) Enter the third attention mechanism module ARF_3, and obtain the feature map through the calculation of ARF_3;

[0045] (8) Enter conv5_x to perform hole convolution calculation;

[0046] (9) Enter the attention mechanism mixing module ARMF for calculation;

[0047] (10) Enter the spatial pyramid module SPP...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The indoor scene semantic annotation method comprises the following steps: inputting a color picture and a depth picture; entering the neural network, wherein the color picture and the depth picture pass through a conv1 and a conv2 _ x respectively; entering a first attention mechanism module ARF _ 1, and obtaining a feature map through calculation of the ARF _ 1; entering conv3 _ x to carry out convolution calculation; entering a second attention mechanism module ARF _ 2, and obtaining a feature map through calculation of the ARF _ 2; entering conv4 _ x to carry out hole convolution calculation; entering a third attention mechanism module ARF _ 3, and obtaining a feature map through calculation of the ARF _ 3; entering conv5 _ x to carry out hole convolution calculation; entering an attention mechanism mixing module ARMF for calculation; entering a space pyramid module SPP to realize multi-level context information fusion; obtaining a semantic annotation result graph.

Description

technical field [0001] The invention relates to the technical fields of multimedia technology and computer graphics, in particular to a semantic labeling method for indoor scenes. Background technique [0002] Scene labeling, or scene parsing, is to label each pixel in the image with the object category label it belongs to. Scene semantic annotation is a challenging task since it combines the traditional problems of detection, segmentation, and multi-label recognition in a single process. High-quality scene annotation is beneficial for intelligent tasks such as robot task planning, pose estimation, plane segmentation, context-based image retrieval, and automatic photo adjustment. [0003] The previous scene identification work can be divided into two categories: indoor scene and outdoor scene according to the target scene. Compared with outdoor scene labeling, indoor scene labeling is more challenging because of the larger set of indoor scene semantic labels, more serious ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62
CPCG06V20/36G06F18/241
Inventor 王立春李玉洁王少帆孔德慧
Owner BEIJING UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products