Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Low-latency collaborative adaptive CNN inference system and low-latency collaborative adaptive CNN inference method

An adaptive, low-latency technology, applied in the field of edge computing and deep learning, which can solve the problems of high synchronization overhead, fixed scheduling strategy, and speeding up CNN inference speed.

Pending Publication Date: 2020-12-29
NANJING UNIV
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] Purpose of the invention: Aiming at the deficiencies of the prior art, the present invention proposes a distributed collaborative adaptive CNN inference system for CNN tasks under the edge network, which at least solves the problems of fixed scheduling strategy and high synchronization overhead, so as to speed up the deployment of CNN at the edge of the network. inference speed

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Low-latency collaborative adaptive CNN inference system and low-latency collaborative adaptive CNN inference method
  • Low-latency collaborative adaptive CNN inference system and low-latency collaborative adaptive CNN inference method
  • Low-latency collaborative adaptive CNN inference system and low-latency collaborative adaptive CNN inference method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0064] The technical solution of the present invention will be further described below in conjunction with the accompanying drawings.

[0065] According to an embodiment of the present invention, a distributed CNN inference system for CNN tasks under the edge network is a collaborative system composed of a single master node and multiple slave nodes, wherein the master node monitors the running status of each slave node in real time, And periodically update slave node tasks to coordinate global tasks. There is only one master node in the system; slave nodes are responsible for inferring the specific execution of tasks, and there are usually multiple slave nodes. Both the master node and the slave node maintain the structural information of CNN, and use the cross-layer data range deduction mechanism to calculate the dependency between any two pieces of data.

[0066] Specifically, refer tofigure 1 , the architecture of the master node on the right, including: running informatio...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a low-latency collaborative adaptive CNN inference system and method, which are applied to an edge computing network. The system comprises a master node and a plurality of slave nodes, the master node is used for monitoring the running state of each slave node in real time and periodically updating inference tasks of the slave nodes, and the slave nodes are used for inferring specific execution of the tasks. According to the invention, a CNN cross-layer data range deduction mechanism is realized in the system, so that the master node and the slave node can deduce the dependency relationship between any two pieces of data in multiple layers of the CNN, thereby realizing efficient user-defined dynamic scheduling; meanwhile, a balanced synchronous scheduler is designedas a built-in scheduler of the system, the load of the slave nodes is adjusted in a dynamic sensing mode, and efficient recovery of a memory is achieved.

Description

technical field [0001] The invention relates to the fields of edge computing and deep learning, in particular to a low-latency cooperative adaptive CNN inference system and method. Background technique [0002] The past decade has witnessed the rise of deep learning. As its typical representative, Convolutional Neural Network (CNN) is widely used in computer vision and video analysis. Using a dedicated CNN, applications can detect and classify objects from images / videos more precisely than before. Despite these advantages, it is worth noting that CNN inference is computationally resource intensive. Taking VGG-16 as an example, classifying an image with a size of 224*224 requires 15.5G multiplication and addition operations. Therefore, traditional solutions tend to transfer images or videos to the cloud with powerful computing power. However, user data is generated at the edge of the network. Long-distance transmission over a WAN introduces latency and jitter that can ad...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F9/48G06F9/50G06N3/04H04L29/08
CPCG06F9/4881G06F9/5016G06F9/5027H04L67/10G06N3/045Y02D10/00
Inventor 张帅张胜钱柱中陆桑璐
Owner NANJING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products