Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

41 results about "Visual modality" patented technology

Visual Modality - A Visual Learner. Learns by seeing and by watching demonstrations Likes visual stimuli such as pictures, slides, graphs, demonstrations, etc. Conjures up the image of a form by seeing it in the “mind’s eye” Often has a vivid imagination.

Image annotation method based on weak matching probability canonical correlation model

The invention discloses an image annotation method and system based on a weak matching probability canonical correlation model, relating to the technical field of processing of network cross-media information. The image annotation method comprises the following steps: obtaining an annotated image and a non-annotated image in an image database, respectively extracting image features and textual features of the annotated image and the non-annotated image, and generating a matched sample set and an unmatched sample set, wherein the matched sample set contains an annotated image feature set and an annotated textual feature set; and the unmatched sample set contains a non-annotated image feature set and a non-annotated textual feature set; training the weak matching probability canonical correlation model according to the matched sample set and the unmatched sample set; and annotating an image to be annotated through the weak matching probability canonical correlation model. According to the invention, correlation between a visual modality and a textual modality is learned by using the annotated image, keywords of the annotated image and the non-annotated image simultaneously; and an unknown image can be accurately annotated.
Owner:INST OF COMPUTING TECH CHINESE ACAD OF SCI +1

Multi-modality vocabulary representing method based on dynamic fusing mechanism

The invention provides a multi-modality vocabulary representing method. The multi-modality vocabulary representing method comprises the steps of calculating a text representing vector of a vocabulary to be represented in text modality and a picture representing vector of the vocabulary to be represented in visual modality; inputting the text representing vector into a pre-established text modality weight model to obtain the weight of the text representing vector in the text modality; inputting the picture representing vector into a pre-established visual modality weight model to obtain the weight of the picture representing vector in picture modality; conducting calculation to obtain a multi-modality vocabulary representing vector according to the text representing vector, the picture representing vector and weights corresponding to the text representing vector and the picture representing vector respectively. The text modality weight model is a neural network model of which input is the text representing vector and output is the weight of the text representing vector in the corresponding text modality; the visual modality weight model is a neural network model of which input is the picture representing vector and output is the weight of the picture representing vector in the corresponding visual modality.
Owner:INST OF AUTOMATION CHINESE ACAD OF SCI

Personalized auricle data model building method based on normal auricle morphological classification

The invention discloses a personalized auricle data model building method based on normal auricle morphological classification. Classification indexes for describing a precise auricle structure and building indexes of the personalized data model for describing a basic auricle shape are included. On one hand, data measuring is carried out on auricle indexes, the position of a crus of helix serves as a boundary for partitioned clustering analysis to obtain the auricle data model; on the other hand, according to personalized index data, the personalized auricle data model is further obtained on the basis of the data model obtained through normal auricle classification. Compared with traditional auricle visual modality description and classification, the data model in the method is obtained after clustering analysis of quantitative auricle morphological indexes, so that classification is more accurate, standard and objective, and real auricle information can be reflected better. The personalized auricle data model built due to the fact that personalized index values are introduced into the auricle data model has significant meaning in manufacturing personalized ear cartilage supports and further reconstructing personalized auricles.
Owner:THE THIRD XIANGYA HOSPITAL OF CENT SOUTH UNIV

Image processing method for glaucoma detection and computer program products thereof

The method comprises storing a set of images captured of the anterior chamber of various eyes, and using a processor for: a) processing some of said stored images by implementing an homogenization process that adjusts an horizontal and a vertical spatial resolution of each image of the set to be the same, and a centering and aligning process that computes statistical properties of the images, and uses said computed statistical properties to compute a centroid and a covariance matrix of each image; b) performing pair-wise distance measures between images of said processed images providing a pair-wise distance matrix; c) analyzing said pair-wise distance matrix by executing a nonlinear dimensionality reduction algorithm that assigns a point in an n-dimensional space associated to each analyzed image; and d) outputting the results of said analysis in a visual way enabling being usable to detect if said eyes suffer from glaucoma.
Owner:UNIV POLITECNICA DE CATALUNYA +1

Display of a single or plurality of picture(s) or visual element(s) as a set or group to visually convey information that otherwise would be typed or written or read or sounded out as words or sentences.

The method for creating or using or re-using one or more visual communication unit(s) to visually portray and convey information to be communicated, visually. This visual communication can be in lieu of or in conjunction with other written, and / or spoken, and / or machine language, voice / audible media, and / or any other communication methods. This is done by creating or using one or more visual element(s), picture(s), geometric object(s), painting(s), drawing(s), video(s), movie(s), clip(s), art(s), animation(s), (please note, these are not an exhaustive list of elements), or re-using already made or existing visual communication unit(s), or a combination of, for visually portraying and conveying the information to be communicated, visually. Then communicating the visual communication unit(s) conveying the information visually, through the communication channel(s), tool(s), format(s), and medium(s) of choice.
Owner:SCHLAKE FARIMEHR

A multimodal vocabulary representation method based on dynamic fusion mechanism

The invention provides a multi-modality vocabulary representing method. The multi-modality vocabulary representing method comprises the steps of calculating a text representing vector of a vocabulary to be represented in text modality and a picture representing vector of the vocabulary to be represented in visual modality; inputting the text representing vector into a pre-established text modality weight model to obtain the weight of the text representing vector in the text modality; inputting the picture representing vector into a pre-established visual modality weight model to obtain the weight of the picture representing vector in picture modality; conducting calculation to obtain a multi-modality vocabulary representing vector according to the text representing vector, the picture representing vector and weights corresponding to the text representing vector and the picture representing vector respectively. The text modality weight model is a neural network model of which input is the text representing vector and output is the weight of the text representing vector in the corresponding text modality; the visual modality weight model is a neural network model of which input is the picture representing vector and output is the weight of the picture representing vector in the corresponding visual modality.
Owner:INST OF AUTOMATION CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products