Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

239 results about "Context vector" patented technology

Context Vectors are created for the context of the target word and also for the glosses of each sense of the target word. Each gloss is considered as a bag of words, where each word has a corresponding Word Vector. These vectors for the words in a gloss are averaged to get a Context Vector corre- sponding to the gloss.

A neural network text classification method based on a multi-knowledge map

The invention relates to a neural network text classification method based on a multi-knowledge map. The method comprises the following steps: inputting a text in a training set into a long-term and short-term memory network to obtain a context vector of the text; extracting entities from each text in the training set, and performing entity matching in the knowledge map; calculating the attentionweights of matched entities and relationships in the knowledge map under context vector to obtain the overall entity vector, the overall relationship vector and the fact triple vector of the text; computing the fact triple vectors under different knowledge maps, calculating the attention weights of these fact triples, obtaining the text representation vectors and inputting them to the full connection layer of the neural network, and using a classifier to calculate the probability of each text belonging to different categories to train the network; using the trained deep neural network model to predict the category of the text to be predicted. This method improves the model's understanding of text semantics and can classify text content more reliably, accurately and robustly.
Owner:FUZHOU UNIV

Method of translation, method for determining target information and related devices

The invention discloses a method for determining target information. The method includes the following steps that encoding is conducted on to-be-processed text information to obtain a source-end vector expression sequence; according to the source-end vector expression sequence, a source-end context vector corresponding to the first moment is obtained, wherein the source-end context vector is used for expressing to-be-processed source-end content; according to the source-end vector expression sequence and the source-end context vector, a first translation vector and / or a second translation vector are / is determined, wherein the first translation vector indicates the source-end content which is not translated in the source-end vector expression sequence within the first moment, and the second translation vector indicates the source-end content which is translated in the source-end vector expression sequence within the second moment; decoding is conducted on the first translation vector and / or a second translation vector and the source-end context vector so as to obtain the target information of the first moment. The invention further provides a method of translation and a device for determining the target information. According to the method for determining the target information, the model training difficulty of a decoder can be reduced, and the translation effect of a translation system is improved.
Owner:SHENZHEN TENCENT COMP SYST CO LTD

Automatic text summarization method based on enhanced semantics

The invention discloses an automatic text summarization method based on enhanced semantics. The method comprises the following steps of: preprocessing a text, arranging words from high to low according to the word frequency information, and converting the words to id; using a single-layer bi-directional LSTM to encode the input sequence and extracting text information features; using a single-layer unidirectional LSTM to decode the encoded text semantic vector to obtain the hidden layer state; calculating a context vector to extract the information, most useful the current output, from the input sequence; after decoding, obtaining the probability distribution of the size of a word list, and adopting a strategy to select summarization words; in the training phase, fusing the semantic similarity between the generated summarization and the source text to calculate the loss, so as to improve the semantic similarity between the summarization and the source text. The invention utilizes the LSTM depth learning model to characterize the text, integrates the semantic relation of the context, enhances the semantic relation between the summarization and the source text, and generates the summarization which is more suitable for the subject idea of the text, and has a wide application prospect.
Owner:SOUTH CHINA UNIV OF TECH

Text simplification method based on word vector query model

The invention provides a text simplification method based on a word vector query model. Based on a sequence and a sequence model, when decoding is conducted, the correlation between the hidden state of a decoder and the word vectors of all vocabularies is obtained through the reference of an attention mechanism to serve as a measurement of the possibility of the words to be generated in the next step. The method includes the following steps that a text encoder is designed, and an original text is compressed; a text simplification decoding generator is designed, and the current hidden layer vector and the context vector at every moment are calculated circularly; the retrieval correlation of each word in a word list is obtained, the predicted words at the current moment are output, and a complete simplified text is obtained; a model for generating the simplified text is trained, and the log likelihood of the predicted words and actual target words is minimized; after training, the complete simplified text is generated. The method can improve the quality and accuracy of the generated text, greatly reduce the number of parameters of the existing method, and reduce the training time andthe memory usage.
Owner:PEKING UNIV

Surveillance system using adaptive spatiotemporal convolution feature representation with dynamic abstraction for video to language translation

A surveillance system is provided that includes an image capture device configured to capture a video sequence of a target area that includes objects and is formed from a set of image frames. The system further includes a processor configured to apply a C3D to the image frames to obtain therefor (i) intermediate feature representations across L convolutional layers and (ii) top-layer features. The processor is further configured to produce a first word of a caption for the sequence by applying the top-layer features to a LSTM. The processor is further configured to produce subsequent words of the caption by (i) dynamically performing spatiotemporal attention and layer attention using the intermediate feature representations to form a context vector, and (ii) applying the LSTM to the context vector, a previous word of the caption, and a hidden state of the LSTM. The system includes a display device for displaying the caption.
Owner:NEC CORP

Named entity recognition model training method and named entity recognition method and device

The invention discloses a named entity recognition model training method, a named entity recognition method and a named entity recognition device. The training method comprises the following steps: preprocessing a corpus sample to obtain a character sequence sample, and labeling a named entity label on the character sequence sample to obtain a training character sequence; pre-training the trainingcharacter sequence based on a first bidirectional language model and a first self-attention mechanism model to obtain a character feature vector and a character weight vector, and fusing the character feature vector and the character weight vector to obtain a second context vector; pre-training the training character sequence based on a second bidirectional language model and a second self-attention mechanism model to obtain a word feature vector and a word weight vector, and fusing the word feature vector and the word weight vector to obtain a second context vector; and training the bidirectional neural network and the conditional random field which are connected in sequence by using the first context vector and the second context vector to obtain a named entity recognition model. According to the method, the training effect of the named entity recognition model is effectively improved, and the named entity recognition accuracy is improved.
Owner:SUNING CLOUD COMPUTING CO LTD

Neural machine translation method by introducing source language block information to encode

The present invention relates to a neural machine translation method for introducing source language block information to encode. The method comprises: inputting bilingual sentence-level parallel data, and carrying out word segmentation on the source language and the target language respectively to obtain bilingual parallel sentence pairs after being subject to word segmentation; encoding the source sentence in the bilingual parallel sentence pairs after being subject to word segmentation according to the time sequence, obtaining the state of each time sequence on the hidden layer of the lastlayer, and segmenting the input source sentence by blocks; according to the state of each time sequence of the source sentence and the segmentation information of the source sentence, obtaining the block encoding information of the source sentence; combing the time sequence encoding information with the block encoding information to obtain final source sentence memory information; and by dynamically querying the source sentence memory information, using attention mechanism to generate a context vector at each moment through a decoder network, and extracting feature vectors for word prediction.According to the method provided by the present invention, block segmentation is automatically carried out on the source sentence without the need of any pre-divided sentence to participate in the training, and the method can capture the latest and the best block segmentation manner of the source sentence.
Owner:沈阳雅译网络技术有限公司

Long text sentiment analysis method fusing Attention mechanism

The invention relates to a long text sentiment analysis method fusing an Attention mechanism. A text sentiment classification model (Bi-Attention model) is built by adopting a bidirectional thresholdrecurrent neural network in combination with the Attention mechanism. Attention can enable a neural network to pay attention to important information in a text and ignore or reduce the influence of secondary information in the text, thereby lowering the complexity of long text processing. On the other hand, a context vector is generated through a bidirectional threshold recurrent unit, and a memory state is updated, thereby fully considering the influence of historical information and future information on semantics.
Owner:FUZHOU UNIV

Method for extracting relationship among geographic entities contained in internet text

The invention discloses a method for extracting a relationship among geographic entities contained in an internet text. The method comprises the following steps of data preprocessing, document vectorization, weight calculation, keyword extraction and relational tuple construction. The method specifically comprises the steps of inputting the network text containing the geometric entities, extracting the spatial relationship or the semantic relationship among the geometric entities through data preprocessing, and obtaining a webpage pure text and candidate keywords; performing vectorization on the text by adopting a word-level vector space model, and establishing a word-context matrix; designing a novel weight calculation method for performing weight calculation on the geometric entities; and selecting the word with the maximum weight as a keyword from a context vector, constructing a relational tuple, and finally finishing geometric entity extraction. According to the method, a semantic-based retrieval mode is provided, so that a conventional search technology depending on the keyword is changed; and on the premise of lack of large-scale tagged corpora and a geometric knowledge library, geometric relationship description words can be quickly extracted, so that the operation efficiency is improved and the labor cost is greatly reduced.
Owner:INST OF GEOGRAPHICAL SCI & NATURAL RESOURCE RES CAS

Representation and retrieval of images using context vectors derived from image information elements

Image features are generated by performing wavelet transformations at sample points on images stored in electronic form. Multiple wavelet transformations at a point are combined to form an image feature vector. A prototypical set of feature vectors, or atoms, is derived from the set of feature vectors to form an “atomic vocabulary.” The prototypical feature vectors are derived using a vector quantization method, e.g., using neural network self-organization techniques, in which a vector quantization network is also generated. The atomic vocabulary is used to define new images. Meaning is established between atoms in the atomic vocabulary. High-dimensional context vectors are assigned to each atom. The context vectors are then trained as a function of the proximity and co-occurrence of each atom to other atoms in the image. After training, the context vectors associated with the atoms that comprise an image are combined to form a summary vector for the image. Images are retrieved using a number of query methods, e.g., images, image portions, vocabulary atoms, index terms. The user's query is converted into a query context vector. A dot product is calculated between the query vector and the summary vectors to locate images having the closest meaning. The invention is also applicable to video or temporally related images, and can also be used in conjunction with other context vector data domains such as text or audio, thereby linking images to such data domains.
Owner:FAIR ISAAC & CO INC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products