Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Method for implementing decussation retrieval between mediums through amalgamating different modality information

A modal, media technology, applied in special data processing applications, instruments, electrical digital data processing, etc., can solve a large number of manual annotation, no way to directly obtain media objects and other problems, to achieve the effect of powerful functions and accurate precision

Inactive Publication Date: 2008-11-12
ZHEJIANG UNIV
View PDF0 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] The development of the Web is accompanied by a sharp increase in the amount of information. Faced with such a huge amount of data, retrieval has become an important means for people to obtain information. Simple text retrieval can no longer meet the increasingly complex needs of users. Just text data also includes data in different modes such as images, videos, audios, slides in Microsoft PowerPoint format
However, the retrieval of existing multimedia objects is generally achieved through manual labeling and matching of underlying features. The multimedia search is only suitable for use in a small area with a limited amount of data; although the matching retrieval based on the underlying features of multimedia objects does not require a lot of manual input, due to the gap between the underlying features and semantics, for example, visually similar images may Represents completely different semantics, and semantically the same image may look completely different, so the retrieval method that integrates the underlying features and semantics is of great significance
[0003] At present, there is no way to directly obtain the semantics of media objects, so we can only make full use of the semantic relationship between media objects to realize retrieval based on semantics and features

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method for implementing decussation retrieval between mediums through amalgamating different modality information
  • Method for implementing decussation retrieval between mediums through amalgamating different modality information
  • Method for implementing decussation retrieval between mediums through amalgamating different modality information

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0064] Assume that there are 1000 hypermedia consisting of 950 images, 100 sound clips and 800 texts. First extract the color features and texture features of all images, where the color features include color histograms, color moments, and color aggregation vectors, and the texture features include roughness, directionality, and contrast, and then calculate the pairwise distance between all images; for sound Segment, extract the Mel frequency cepstral coefficient MFCC, and calculate the distance between all sound objects; for text, calculate the distance between two text objects after vectorization of lexical frequency / inverse document frequency. After the media object distance calculation is completed, the image distance, text distance and sound distance should be normalized respectively. Establish audio distance map A, image distance map I, and text distance map T between hypermedia objects. To establish audio distance map A, firstly, for any hypermedia objects A and B, fir...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a method for realizing the interaction searching between medias by integrating different modal information which includes the following steps: 1) building relation graphs to hypermedias and obtaining corresponding relative coefficient matrixes; 2) using media objects or the hypermedias inside or outside a database submitted by a user as a searching example to mark the initial matching degree; 3) utilizing the relative coefficient matrix between the hypermedias to repeatedly iterate to a stable state and broadcasting the matching ability to the un-marked hypermedias and returning to the hypermedias or the media objects of special modal states in the hypermedias with the matching degree of the searching example larger than 0.6; 4) periodically carrying out adjustment on the relation graphs of the hypermedia according to a searching example set and a positive example set. The invention integrates the bottom layer characteristics of various media objects and broadcasts the semanteme by the sibship between the media objects, thereby having a better searching effect; as the searching example and the state of a returning result can be different and are broadcasted by utilizing the semanteme, the searching is more accurate and the adaptation is broader.

Description

technical field [0001] The present invention relates to cross-retrieval of media between different modalities, in particular to a method for merging information of different modalities to realize cross-retrieval between media. Background technique [0002] The development of the Web is accompanied by a sharp increase in the amount of information. Faced with such a huge amount of data, retrieval has become an important means for people to obtain information. Simple text retrieval can no longer meet the increasingly complex needs of users. Just text data also includes data of different modalities such as images, videos, audios, and slides in Microsoft PowerPoint format. However, the retrieval of existing multimedia objects is generally achieved through manual labeling and matching of underlying features. The multimedia search is only suitable for use in a small area with a limited amount of data; although the matching retrieval based on the underlying features of multimedia o...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30
Inventor 吴飞庄越挺王文华杨易
Owner ZHEJIANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products