Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Cross-modal retrieval method based on adversarial learning and asymmetric hashing

A cross-modal and asymmetric technology, applied in the field of cross-modal retrieval, which can solve the problem that the hash code is not optimal, and the modal data cannot be fully extracted.

Active Publication Date: 2019-09-10
INST OF INFORMATION ENG CAS
View PDF11 Cites 28 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the existing cross-modal methods based on deep learning cannot fully extract the characteristics of each modal data in the feature extraction stage, and the hash codes generated in the hash code generation stage are not optimal.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing
  • Cross-modal retrieval method based on adversarial learning and asymmetric hashing

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0051] In order to better express the cross-modal retrieval method based on adversarial learning and asymmetric hashing proposed in the present invention, the following will take a 224×224 pixel picture and its corresponding text description as an example in combination with the accompanying drawings and specific implementation methods , the present invention will be further described.

[0052] figure 1 It is the overall flowchart of the present invention, including five parts: data preprocessing, initializing model framework, model training, generation of hash codes for each modal data, and retrieval stages.

[0053] Step 1. Data preprocessing. Divide the cross-modal data set into two parts: training set and test set, each data instance contains image-text pairs and their corresponding labels;

[0054] Step 2. Initialize the model framework. figure 2 is the model framework designed in the present invention, which includes a cross-modal feature extraction module, an attent...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a cross-modal retrieval method based on adversarial learning and asymmetric hashing. The cross-modal retrieval method comprises the following steps: 1) selecting a cross-modaldata set and dividing the cross-modal data set into a training set and a test set; training a model, wherein the model comprises a loss module, a feature extraction unit and a Hash code learning unit,wherein the hash code learning unit comprises a hash code learning network and a shared multi-label binary matrix, and the loss module comprises the confrontation loss of the feature extraction unit,the ternary margin loss of the hash code learning unit, the cosine quantization loss and the asymmetric hash loss; 2) generating a binary hash code of each modal data in the cross-modal data set by using the trained optimal model; and 3) for a given query data, firstly generating a binary hash code, and then calculating a Hamming distance between the binary hash code of the query data and the binary hash code of the modal data in the cross-modal data set, which is different from the modal of the query data, so as to obtain a cross-modal data instance meeting the condition.

Description

technical field [0001] The invention relates to a cross-modal retrieval method based on adversarial learning and asymmetric hashing, and belongs to the technical field of computer software. Background technique [0002] With the massive increase of multimedia data in online social media and search engines, there is an increasing demand for mutual retrieval between different modal data (such as video, image, text, etc.). For example, when a user inputs a text query, it needs to obtain picture information or audio and video information related to the query text. In order to solve the retrieval problem in large-scale cross-modal data, an efficient and fast method is the hash method. The cross-modal hash method mainly has two stages, one is the extraction of the characteristics of each modal data, and the other is It is the similarity mining between various modal data. Currently existing cross-modal hashing methods can be roughly divided into two types: traditional cross-modal...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/31G06F16/335G06F16/58G06F16/583G06F16/51G06N3/04G06N3/08
CPCG06F16/325G06F16/335G06F16/5846G06F16/5866G06F16/51G06N3/08G06N3/045
Inventor 古文李波古晓艳熊智谷井子王伟平
Owner INST OF INFORMATION ENG CAS
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products