Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Convolution neural network parallel processing method based on large-scale high-performance cluster

A convolutional neural network and parallel processing technology, which is applied in the field of convolutional neural network parallel processing, can solve problems such as network model size limitation and single server configuration, and achieve the effects of optimizing computing efficiency, improving the scope of application, and improving computing efficiency

Inactive Publication Date: 2015-03-25
CHANGSHA MASHA ELECTRONICS TECH
View PDF3 Cites 157 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] At present, the international cutting-edge research on the acceleration of neural networks mainly focuses on two directions: first, parallel acceleration based on multiple GPUs on a single server, a single server does not involve data transmission between multiple nodes, parallel acceleration is easy to achieve, but The scale of the network model is limited by the configuration of a single server; second, using a large-scale cluster to accelerate the training of the neural network, the DistBelief model was proposed, but it was not applied to the convolutional neural network. In the restricted Boltz Mann machine and deep belief network are widely used

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Convolution neural network parallel processing method based on large-scale high-performance cluster
  • Convolution neural network parallel processing method based on large-scale high-performance cluster
  • Convolution neural network parallel processing method based on large-scale high-performance cluster

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0033] The present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments.

[0034] When the present invention is specifically applied, it is first necessary to construct a large-scale high-performance cluster environment. The high-performance cluster environment is divided into software environment and hardware environment. The hardware is a group of 1 to N independent computing (nodes) with the same configuration. The nodes are connected through a high-performance Internet; each node can be used as a single computing resource for interactive users, and also can work together and appear as a single, centralized computing resource for use by parallel computing tasks.

[0035] The tasks of high-performance clusters are mainly focused on scientific computing, so the requirements for hardware computing capabilities are relatively high. In addition to choosing a CPU with a higher frequency and more cores, the GP...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a convolution neural network parallel processing method based on a large-scale high-performance cluster. The method comprises the steps that (1) a plurality of copies are constructed for a network model to be trained, model parameters of all the copies are identical, the number of the copies is identical with the number of nodes of the high-performance cluster, each node is provided with one model copy, one node is selected to serve as a main node, and the main node is responsible for broadcasting and collecting the model parameters; (2) a training set is divided into a plurality of subsets, the training subsets are issued to the rest of sub nodes except the main mode each time to conduct parameter gradient calculation together, gradient values are accumulated, the accumulated value is used for updating the model parameters of the main node, and the updated model parameters are broadcast to all the sub nodes until model training is ended. The convolution neural network parallel processing method has the advantages of being capable of achieving parallelization, improving the efficiency of model training, shortening the training time and the like.

Description

technical field [0001] The invention mainly relates to the design field of high-performance computing clusters, in particular to a convolutional neural network parallel processing method based on large-scale high-performance clusters. Background technique [0002] A high-performance computer is a computer cluster system, which connects multiple computer systems together through high-speed interconnection technology, and uses the comprehensive computing power of all connected systems to deal with large-scale computing problems, so it is often called a "high-performance computing cluster". Or "high-performance clusters". High-performance clusters are mainly used to deal with complex computing problems and are used in environments that require large-scale scientific computing, such as weather forecasting, oil exploration and reservoir simulation, molecular simulation, gene sequencing, etc. Applications running on high-performance clusters generally use parallel algorithms to d...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/02G06F9/38G06F15/16
Inventor 王馨
Owner CHANGSHA MASHA ELECTRONICS TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products