Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A fast spectral clustering method based on improved kd-tree marker selection

A technology of marking points and spectral clustering, which is applied in special data processing applications, instruments, electrical digital data processing, etc., and can solve problems such as network expansion

Inactive Publication Date: 2019-02-01
LIAONING TECHNICAL UNIVERSITY
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

But the size of the input layer of the neural network is equal to the number of data points, as n grows, the entire network will expand dramatically

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A fast spectral clustering method based on improved kd-tree marker selection
  • A fast spectral clustering method based on improved kd-tree marker selection
  • A fast spectral clustering method based on improved kd-tree marker selection

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0033] The preferred implementation of the method for fast spectral clustering based on the improved kd tree marker point selection of the present invention will be described in detail below.

[0034] Input: a dataset with n data points

[0035] Output: Divide the dataset into k classes

[0036] 1. Use the improved KD tree to select marker points to select p marker points from X, denoted as

[0037] 2. Calculate the similarity between all data points and marked points, and store them in the matrix W.

[0038] 3. Calculate degree matrix

[0039] 4. Calculate the input S of the self-encoder,

[0040] 5. Use S as input to train the autoencoder

[0041] 6. Perform k-means clustering in the hidden layer of the trained autoencoder.

[0042] Self-encoding instructions used

[0043] The input of the self-encoder in this paper is S=WD -1 / 2 . The objective function for training the autoencoder is the reconstruction error of S. After training the autoencoder, we obtain the repre...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a fast spectrum clustering method based on improved kd tree mark point selection. The invention firstly uses the improved kd tree method to select the marker points in the dataset, then establishes the similarity matrix between the data according to the marker points, and finally uses a depth self-coding instead of the traditional dimension reduction method for eigendecomposing the Laplace matrix. In the experiment, we prove the superiority of our method. In the large-scale data set, the algorithm improves the speed of clustering greatly with the accuracy basically consistent with the classical spectral clustering algorithm.

Description

technical field [0001] The invention relates to the field of clustering under data mining, in particular to a fast spectral clustering method based on the selection of improved kd tree marker points. Background technique [0002] Spectral clustering comes from the minimum cut problem in graph theory. If the data set is represented as a graph, each data point is used as a vertex of the graph, and the edges connecting the vertices represent the similarity between adjacent vertices. Then the process of spectral clustering is to divide the similarity graph into several subgraphs by cutting the edge with the smallest similarity, so that the similarity in the subgraph is the largest and the similarity between the subgraphs is the smallest. [0003] Spectral clustering works very well for high-dimensional small-scale data clustering because of its unique step of decomposing the Laplacian matrix into features. However, the computational complexity of this eigen-decomposition step ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/901G06F16/2458
Inventor 邱云飞王秋涵
Owner LIAONING TECHNICAL UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products