Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-mode 3D target detection method and system thereof, terminal and medium

A target detection, multi-modal technology, applied in the field of 3D target detection, can solve problems such as inapplicability, no description or report found, inaccurate prediction of object edge depth, etc., to achieve high-quality detection and high-performance improvement.

Inactive Publication Date: 2021-06-18
SHANGHAI JIAO TONG UNIV
View PDF11 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0007] 1. The input of this method is an RGB-D image, which is not suitable for sparse point clouds in outdoor scenes
[0008] 2. In this method, the extraction of the candidate frame of the RGB image is decoupled from the branch of the 3D point cloud, and the classification of the 3D point cloud is also decoupled from the RGB image, so the complementary information between the image and the point cloud cannot be fully utilized.
[0010] 1. The feature fusion of point cloud data and RGB image input in this method only stays at the level of feature splicing, and does not consider the depth correlation between the two modal attributes
[0011] 2. The supervision signal of the detection result correction network of this method comes from the completed point cloud data obtained by completing the point cloud data based on the fusion data. However, due to the inaccurate prediction of the depth of the edge of the object, the completed point cloud will be Introduce some noise to the correction network, thereby weakening the performance of the correction network
[0012] At present, there is no description or report of the similar technology of the present invention, and no similar data at home and abroad have been collected yet.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-mode 3D target detection method and system thereof, terminal and medium
  • Multi-mode 3D target detection method and system thereof, terminal and medium
  • Multi-mode 3D target detection method and system thereof, terminal and medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0056] The following is a detailed description of the embodiments of the present invention: this embodiment is implemented on the premise of the technical solution of the present invention, and provides detailed implementation methods and specific operation processes. It should be noted that those skilled in the art can make several modifications and improvements without departing from the concept of the present invention, and these all belong to the protection scope of the present invention.

[0057] figure 1 It is a flowchart of a multimodal 3D object detection method provided by an embodiment of the present invention.

[0058] The multi-modal 3D object detection method provided by this embodiment first uses the image and the original point cloud as input, respectively inputs the independent feature extractor to extract the corresponding features, and then performs the point-pixel level feature fusion in the first stage to generate high-quality 3D region proposal. Then the...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a multi-modal 3D target detection method and a system thereof. The method comprises the following steps: respectively extracting features of an original image I and a corresponding LiDAR point cloud L; carrying out point and pixel feature fusion on the features of the original image I and the corresponding LiDAR point cloud L to form LiDAR point cloud features, taking the features of the original image I as image features, and respectively generating a 3D region proposal and a 2D region proposal; extracting features from the 3D area proposal and the 2D area proposal respectively and fusing, and generating a final 3D target detection result. Meanwhile, the invention provides a corresponding terminal and a medium. According to the method, target detection is completed by using a geometric constraint relationship and feature relevance between modals; a 3D target detection task is completed through feature fusion of a point-pixel level of a first stage and a region proposal level of a second stage; and a high-quality regional proposal is generated by using geometric constraints between the image and the LiDAR point cloud.

Description

technical field [0001] The present invention relates to a 3D target detection method, in particular to a deep network-based multi-modal 3D target detection method, system, terminal and medium. Background technique [0002] Object detection is an important direction in the field of computer vision, and has broad application prospects and market value. With the development of various sensor technologies and automotive technologies, the role of 3D object detection in the field of autonomous driving has gradually emerged. The most commonly used sensors for 3D object detection tasks in autonomous driving include cameras and LiDAR, and the corresponding data types are images and LiDAR point clouds, respectively. Due to the complementarity of information between the two modalities of image and LiDAR point cloud, and the gradually decreasing price of LiDAR, 3D target detection methods based on multimodality have gradually become the focus of research at home and abroad. [0003] E...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/46G06K9/62
CPCG06V10/40G06V2201/07G06F18/24G06F18/253G06F18/214
Inventor 马超
Owner SHANGHAI JIAO TONG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products