Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Visual feature representing method based on autoencoder word bag

An autoencoder and visual feature technology, applied in neural learning methods, instruments, biological neural network models, etc., can solve problems such as overfitting, difficulty in learning effective features, and difficulty in discovering data structures. Applicability, practicality, simple and feasible effect

Active Publication Date: 2014-12-24
TIANJIN UNIV
View PDF5 Cites 25 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Due to the inherent properties of the deep architecture (contains more hidden layers) and the use of the entire image as input, SAE is only suitable for situations where the image size is small and the number of training samples is large, while the image size is relatively large and the number of samples In rare cases, deep networks (not just SAE) are prone to overfitting, making it difficult to learn effective features
[0006] In the feature representation method based on deep network, because the deep network contains a large number of parameters, it is easy to overfit when there are few training samples; in the feature representation method based on visual bag of words, SIFT and other artificially designed characteristics, making it difficult to discover the inherent structure of the data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Visual feature representing method based on autoencoder word bag
  • Visual feature representing method based on autoencoder word bag
  • Visual feature representing method based on autoencoder word bag

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0035] A method for representing visual features based on an autoencoder bag-of-words of the present invention will be described in detail below with reference to the embodiments and the accompanying drawings.

[0036] A visual feature representation method based on an autoencoder bag of words of the present invention is mainly based on a traditional feature representation method based on a visual bag of words, and uses an autoencoder in unsupervised learning to extract local features of an image for the first time. To replace the manually designed feature, the present invention calls the visual feature representation method based on the autoencoder word bag Bag of Autoencoder Words (BOAEW). BOAEW combines the bag-of-visual-words framework and unsupervised feature learning methods to improve feature representation capabilities.

[0037] In supervised learning, the training samples are labeled with categories. In contrast, in unsupervised learning, the training set has no categ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a visual feature representing method based on an autoencoder word bag. The method includes the steps that training samples are input to form a training set; the training samples in the training set are preprocessed, and influences of illumination, noise and the like on the image representing accuracy are reduced; a visual dictionary is generated, an autoencoder is used for extracting random image block features, then the clustering method is used for clustering the random image block features into a plurality of visual words, and the visual dictionary is composed of the visual words; a sliding window mode is used for sequentially collecting image blocks of images in the training set, the collected image blocks serve as input of the autoencoder, and output of the autoencoder is local features of the images; the local features of the images are quantized into the visual words according to the visual dictionary; the frequency of the visual words is counted, a visual word column diagram is generated, and the visual word column diagram is overall visual feature representing of the images. By means of the visual feature representing method, feature representing characteristics are independently studied through the adoption of the autoencoder, and requirements for the quantity of the training samples are reduced through a BoVW framework.

Description

technical field [0001] The invention relates to a visual feature representation method in the fields of multimedia analysis, machine vision and the like. In particular, it involves a visual feature representation method based on autoencoder bag-of-words. Background technique [0002] Visual representation is one of the important issues in the fields of multimedia analysis, machine vision, etc. It profoundly affects many practical application problems such as multimedia retrieval, image classification, scene analysis, and object recognition. In recent years, with the rapid development of network technology, more and more digital images have flooded people's lives. How to accurately represent these visual data (especially in the environment of big data) has become an urgent problem to be solved. [0003] The bag of words (Bag of Words, BoW) model was originally applied in the field of document retrieval and classification. Its basic idea is: all the words in the statistical c...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62G06N3/08
Inventor 冀中刘青
Owner TIANJIN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products