Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-modal image-text recommendation method and device based on deep learning

A technology of deep learning and recommendation methods, which is applied in the field of computer science and technology applications, can solve problems such as insufficient retrieval accuracy and long time-consuming, and achieve the effects of improving quality and efficiency, improving efficiency and accuracy, and fast training speed

Active Publication Date: 2021-07-09
SHAANXI NORMAL UNIV
View PDF11 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

At present, many cross-modal retrieval methods are only applied in the field of retrieval, and have not been applied to the field of recommendation system.
Moreover, these cross-modal retrieval methods have shortcomings such as insufficient retrieval accuracy and long time consumption.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-modal image-text recommendation method and device based on deep learning
  • Multi-modal image-text recommendation method and device based on deep learning
  • Multi-modal image-text recommendation method and device based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0062] The technical solutions of the present invention will be further described below in conjunction with the accompanying drawings and examples.

[0063] Using a cross-modal image-text retrieval model MMDNN (Multimodal Deep Neural Network), MMDNN is used in the recommendation system, using the positive and negative feedback clustering center calculation module PNFCCCM (Positive and Negative Feedback Cluster Center Calculation Module) and the user's positive and negative feedback Historical records, calculate the user's positive and negative feedback clustering centers; combine the data similarity score and positive and negative feedback scores, find out from the database the data with the highest comprehensive score in the user's historical records. Then use the MMDNN model to find another mode of data corresponding to these pieces of data from the database. Finally, the paired graphic-text resources are recommended to the user, and the user's history and the user's positiv...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-modal image-text recommendation method and device based on deep learning. The method comprises the following steps: using a cross-modal image-text retrieval model MMDNN; secondly, the MMDNN is used in a recommendation system, and a positive and negative feedback clustering center of the user is calculated by using a positive and negative feedback clustering center calculation module PNFCCCM and positive and negative feedback historical records of the user; in combination with the similarity score and the positive and negative feedback score of the data, searching several pieces of data with the highest comprehensive score in the historical records of the user are from the database, and using the MMDNN model for finding out data of another mode corresponding to the several pieces of data from the database, and finally, recommending paired graph-text resources to the user; and updating the historical record of the user and the positive and negative feedback clustering center of the user according to the feedback of the user, so that multi-modal image-text recommendation is realized.

Description

technical field [0001] The invention belongs to the field of computer science and technology application, and in particular relates to a deep learning-based multimodal graphic-text recommendation method and device. Background technique [0002] Currently, most recommender systems focus on providing a single mode of content, such as using images to recommend images and using text to recommend text. In fact, images and text resources in different forms have an unbalanced and complementary relationship when describing the same semantics. Images can usually contain more details that cannot be displayed in text, and text has the advantage of expressing high-level meaning. Therefore, users need more multimodal information resources, and cross-modal retrieval technology is more concerned. Cross-modal retrieval is a technology that can return information combined with multiple modes according to the information entered by the user in one mode. At present, many cross-modal retrieva...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/535G06F16/335G06K9/62G06N3/08
CPCG06F16/535G06F16/335G06N3/08G06F18/23213G06F18/241G06F18/25
Inventor 黄昭胡浩武
Owner SHAANXI NORMAL UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products