Cross-modal retrieval method and system based on unmarked data migration

A cross-modal, unlabeled technology, applied in other database retrieval, other database query, special data processing applications, etc., can solve the problems of heterogeneity, unlabeled data, insufficient training data, etc., to overcome heterogeneity Effects of differences, improving accuracy, and increasing similarity

Active Publication Date: 2020-01-03
INST OF INFORMATION ENG CAS
View PDF6 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In order to overcome the problems of data heterogeneity in different modalities, excessive unlabeled data, insufficient training data and non-extensibility, the present invention proposes a cross-modal retrieval method and system based on unlabeled data migration. The single-modal image and text data with labeled information are used as the migration source domain, and the cross-modal data set with labeled information is used as the target domain. Through transfer learning, the source domain is migrated to the cross-modal data set of the target domain to expand the training data scale. Increase semantic information across modal data to learn a better common space

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cross-modal retrieval method and system based on unmarked data migration
  • Cross-modal retrieval method and system based on unmarked data migration
  • Cross-modal retrieval method and system based on unmarked data migration

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0065] During the concrete implementation of the present invention, comprise three parts of training system, feature extraction system and retrieval: above-mentioned three modules are combined to be general structure of the present invention ( figure 1 ), transfer the training data to the training system for training, and save the training model. feature extraction system ( image 3 ) with the same parameters as the training system, but without the need for data migration, category word embedding and other structures, the test set is passed into the feature extraction system, and the vector representation of each sample in the test set is obtained. When retrieving, calculate the distance between the sample to be retrieved and all samples of other modalities, and the ones that are less than the specified threshold are the retrieval results.

[0066] Training system:

[0067] Such as figure 1 As shown, the combination of the above three modules (unlabeled data clustering modu...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a cross-modal retrieval method and system based on unmarked data migration. The clustered single-mode image and text data without annotation information are used as a migrationsource domain, a cross-mode data set with annotation information is used as a target domain, the source domain is migrated to the cross-mode data set of the target domain through migration learning, the training data scale is expanded, the semantic information of the cross-mode data is increased, and a better common space is learned. According to the method and system, the problem that the cross-modal data set is small in data scale is well solved, and the condition that actual user query is not within a predefined category range is better met; and meanwhile, the upper-layer semantic information of different modal data can be better extracted, the isomerism difference between modals is overcome, the similarity between the modals is increased, and the cross-modal retrieval accuracy is improved.

Description

technical field [0001] The invention relates to the technical field of cross-modal data retrieval, in particular to a cross-modal retrieval method and system based on unmarked data migration. Background technique [0002] Data of different modalities, such as images and texts, widely exist on the Internet and show a trend of mutual integration. The cross-modal retrieval task attempts to break the boundaries between different modal data, and achieve information retrieval across different modal data, that is, try to use a certain modal sample to retrieve samples of other modals with similar semantics. It is widely used in data management. Existing cross-modal retrieval methods try to map the feature representations of different modal data into a common space to learn a unified representation, and measure the similarity by calculating the distance between their corresponding unified representations. However, due to the heterogeneity of different modal data, data distribution ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/62G06F16/903
CPCG06F16/903G06F18/23211G06F18/214
Inventor 朱福庆王雪如张卫博戴娇虎嵩林韩冀中
Owner INST OF INFORMATION ENG CAS
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products