Eureka AIR delivers breakthrough ideas for toughest innovation challenges, trusted by R&D personnel around the world.

A road scene semantic segmentation method based on a convolution neural network

A convolutional neural network and semantic segmentation technology, applied in the field of semantic segmentation of deep learning, can solve the problems of image feature information reduction, rough restoration effect information, and single feature map

Active Publication Date: 2019-03-08
ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY
View PDF2 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] Most of the existing road scene semantic segmentation methods use deep learning methods. There are many models that combine convolutional layers and pooling layers. However, the feature maps obtained by purely using pooling operations and convolution operations are single and not representative. , which will lead to the reduction of the feature information of the obtained image, which will eventually lead to rough restoration of the effect information and low segmentation accuracy.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A road scene semantic segmentation method based on a convolution neural network
  • A road scene semantic segmentation method based on a convolution neural network
  • A road scene semantic segmentation method based on a convolution neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0051] The present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments.

[0052] A method for road scene semantic segmentation based on convolutional neural network proposed by the present invention, its overall realization block diagram is as follows figure 1 As shown, it includes two processes of training phase and testing phase;

[0053] The specific steps of the described training phase process are:

[0054] Step 1_1: Select Q original road scene images and the real semantic segmentation images corresponding to each original road scene image, and form a training set, and record the qth original road scene image in the training set as {I q (i,j)}, combine the training set with {I q (i, j)} corresponding to the real semantic segmentation image is denoted as Then, the existing one-hot encoding technology (one-hot) is used to process the real semantic segmentation images corresponding to each original road scene...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a road scene semantic segmentation method based on a convolution neural network. In a training stage, a convolution neural network is constructed. The hidden layer comprises five neural network blocks, five transition convolution layers, five skip deconvolution blocks and four cascade layers. The original road scene images are inputted into the convolution neural network for training, and 12 corresponding semantic segmentation prediction maps are obtained. Secondly, by calculating the loss function value between the set of 12 semantic segmentation prediction images corresponding to the original road scene images and the set of 12 heat-coded images corresponding to the real semantic segmentation images, the optimal weight vector and bias term of the classification training model of the convolution neural network are obtained. In the testing phase, the road scene images to be semantically segmented are inputted into the convolution neural network classification training model to obtain the predictive semantic segmentation images. The invention has the advantages of improving the efficiency and accuracy of the semantic segmentation of the road scene images.

Description

technical field [0001] The invention relates to a deep learning semantic segmentation method, in particular to a convolutional neural network-based semantic segmentation method for road scenes. Background technique [0002] The rise of the intelligent transportation industry has led to more and more applications of semantic segmentation in intelligent transportation systems. From traffic scene understanding and multi-target obstacle detection to visual navigation, semantic segmentation technology can be used to achieve. Currently, the most commonly used semantic segmentation methods include algorithms such as support vector machines and random forests. These algorithms mainly focus on binary classification tasks to detect and recognize specific objects such as road surfaces, vehicles, and pedestrians. These traditional machine learning methods often need to be implemented through high-complexity features, but it is simple and convenient to use deep learning to semantically ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/34G06N3/04G06N3/08
CPCG06N3/08G06V20/35G06V10/267G06N3/045
Inventor 周武杰顾鹏笠潘婷吕思嘉钱亚冠向坚
Owner ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Eureka Blog
Learn More
PatSnap group products