Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A Feature Fusion Method for Multimodal Deep Neural Networks

A deep neural network and feature fusion technology, applied in the feature fusion field of multi-modal deep neural network, can solve the problem of feature weight distribution analysis and processing, and achieve the effect of maximizing performance

Active Publication Date: 2021-03-30
ZHEJIANG LAB +1
View PDF3 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, these methods did not analyze and process the feature weight distribution of each mode, but simply added, superimposed or fused directly

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Feature Fusion Method for Multimodal Deep Neural Networks
  • A Feature Fusion Method for Multimodal Deep Neural Networks

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0016] Taking the PET / CT dual mode as an example (that is, when x=2), the present invention will be described in detail in conjunction with the accompanying drawings.

[0017] Such as figure 1 Shown, the inventive method specifically comprises the following steps:

[0018] Step 1: In the dual-branch dual-modal 3D CNN, the two branches correspond to the convolution branch of the PET modality and the convolution branch of the CT modality respectively. For the 3D feature maps output by the nth level of the two 3D convolution branches, the 3D feature maps of the two modalities are superimposed on the channel dimension to obtain a 3D feature map with twice the number of original channels. Then perform average pooling in the three dimensions of depth, height, and width, and compress the three dimensions of depth, height, and width to obtain a one-dimensional vector of a channel dimension. After downsampling and upsampling with a compression ratio of 16:1:16, and using the activati...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a feature fusion method of a multi-modal deep neural network. In the multi-modal deep three-dimensional CNN, by using a compression excitation (squeeze and excitation, S&E) module on the deep learning feature domain, the information about the mode can be obtained. Channel attention mask between modalities, that is, in all modalities, give greater attention to those channels that significantly help the task goal, thus explicitly establishing the weight of the multimodal 3D depth feature map on the channel distribution; then, using four-dimensional convolution and Sigmoid activation function calculation, the spatial attention mask between the modalities can be obtained, that is, in the three-dimensional feature map of each modality, which positions in space need to be given greater attention , thereby explicitly establishing the spatial correlation of the multimodal 3D depth feature map, and giving greater attention to the positions with important information in the modality, channel, and space, thereby improving the diagnosis of the multimodal intelligent diagnosis system efficacy.

Description

technical field [0001] The invention relates to the fields of medical imaging and deep learning, in particular to a feature fusion method of a multimodal deep neural network. Background technique [0002] Existing tumor detection and diagnosis methods are usually realized through medical imaging technology, including planar X-ray imaging, CT, MRI, PET / CT, ultrasound and other modalities, and tissue biopsy is performed on suspicious lesions found in the images. However, due to the heterogeneity of tumors, their properties cannot be fully characterized by a single modality image. For example, on plane X-ray and CT images, the X-ray absorption degree of tumor tissue is characterized; on MRI images, the hydrogen proton density of tumor tissue is characterized; on FDG PET / CT, the It is the activity of tumor tissue to metabolize glucose; on ultrasound images, it is characterized by the reflection degree of tumor tissue to sound waves. Therefore, more and more clinical research b...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/62G06N3/04
CPCG06N3/045G06F18/253
Inventor 陈凌朱闻韬张铎申慧李辉叶宏伟王瑶法
Owner ZHEJIANG LAB
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products