Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multistage target detection method and model based on CNN multistage feature fusion

A feature fusion and target detection technology, applied in the field of computational vision target detection, can solve the problems of increasing the target area of ​​interest, increasing complexity, loss of feature information and position information, etc., to optimize the network structure, improve accuracy, and improve accuracy degree of effect

Active Publication Date: 2018-09-07
CENT SOUTH UNIV
View PDF14 Cites 119 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, these classic target detection techniques generally have deficiencies: the targets in the image are often diverse in terms of pose, scale, aspect ratio, etc., and cannot detect multiple categories of targets of different sizes well, especially in complex scenes. When the target scale is changeable and the target scale is relatively small; because these model structures have the characteristics of hierarchical convolution downsampling, the feature information and position information extracted for some relatively small-scale targets are often lost, resulting in some high-semantic information of the target. The consequences of accurate positioning; in addition, the accuracy and efficiency of general target detection are still not well balanced
This method has the following shortcomings: (1) All the features of the convolutional layer are fused, and the relationship between the target size in the image and the high and low features output by the convolutional layer is not considered, that is, the low layer with high resolution is excessively combined Features and high-level features with high semantic information will increase unnecessary computational complexity; (2) The way of feature fusion is the key to the performance of small target detection, but it does not give the connection of multi-layer features to be fused The method is only to connect the output size with the output feature size of a certain convolutional layer; (3) This scheme does not provide a detection network model with appropriate speed and high accuracy for applying its method
However, the biggest problem with this method is that too many target regions of interest are added, resulting in too many irrelevant segment features, increasing the complexity, and not distinguishing the detection of different sizes of targets in the image. If the image contains a large number of relatively large The target increases the calculation amount of the target detection

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multistage target detection method and model based on CNN multistage feature fusion
  • Multistage target detection method and model based on CNN multistage feature fusion
  • Multistage target detection method and model based on CNN multistage feature fusion

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0054] The main idea of ​​the present invention is to fully consider the relationship between the scale size of the target in the image and the high-level and low-level feature maps, and further improve the detection of different-sized targets on the basis of balancing the speed and accuracy of target detection, so as to improve the detection of multiple types of targets. overall detection performance.

[0055] In order to make the technical solution of the present invention clearer and easier to understand, the present invention will be further described below in conjunction with the accompanying drawings and specific embodiments.

[0056] see figure 1 , the present invention provides the detection of objects of different sizes in the high-level and low-level feature maps in the image. In the existing general detection network, the target candidate frame is only extracted from the last layer of feature maps (high-level feature maps), such as figure 1 As shown in (a), when th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-stage target detection method and model based on CNN multistage characteristic fusion; the method mainly comprises the following steps: preparing a related image data set, and processing the data; building a base convolution nerve network (BaseNet) and Feature-fused Net model; training the network model built in the previous step so as to obtain a model of the corresponding weight parameter; using a special data set to finely adjust the trained detection model; outputting a target detection model so as to make target classification and identification, and providing a detection target frame and the corresponding precision. In addition, the invention also provides a multi-class target detection structure model based on CNN multistage characteristic fusion, thus improving the whole detection accuracy, optimizing model parameter values, and providing a more reasonable model structure.

Description

technical field [0001] The invention relates to the technical field of computational vision target detection, in particular to a CNN-based multi-level feature fusion multi-class target detection method and model. Background technique [0002] Target detection is a basic and very important research topic in the field of computational vision, involving image processing, machine learning, pattern recognition and many other disciplines. Its task is to classify and detect corresponding targets from images or videos to be processed. , and provide the specific position and accuracy information of the detected target. With the in-depth research and innovation of this technology, it has been widely used in automotive autonomous driving, video surveillance and analysis, face recognition, vehicle tracking and traffic flow statistics; and target detection is the basis for subsequent image analysis, understanding and application. Therefore, it has important research significance and app...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V2201/07G06N3/045G06F18/253
Inventor 谭冠政刘西亚陈佳庆赵志祥
Owner CENT SOUTH UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products