Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Adjustable hardware aware pruning and mapping framework based on ReRAM neural network accelerator

A neural network and mapping framework technology, applied in the field of computer science artificial intelligence, can solve the problems of inefficient model mapping performance, restricting the development of ReRAM neural network accelerator performance advantages, etc.

Inactive Publication Date: 2021-04-02
ZHEJIANG LAB +1
View PDF3 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, when the hardware specifications and types of ReRAM neural network accelerators are different, and when faced with different levels of user requirements such as delay and energy consumption, the traditional deep learning pruning scheme cannot perceive the changes in hardware and user requirements and make the same A pruning scheme, resulting in the performance inefficiency of model mapping on the ReRAM neural network accelerator hardware, which restricts the development of the performance advantages of the ReRAM neural network accelerator

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Adjustable hardware aware pruning and mapping framework based on ReRAM neural network accelerator
  • Adjustable hardware aware pruning and mapping framework based on ReRAM neural network accelerator
  • Adjustable hardware aware pruning and mapping framework based on ReRAM neural network accelerator

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0021] figure 1 It is the overall structural diagram of the hardware-aware pruning and mapping framework of the present invention, as shown in the figure, including a DDPG agent and a ReRAM neural network accelerator; wherein, the ReRAM neural network accelerator includes multiple processing units (ProcessingElement, PE), each processing The unit is composed of a cross array composed of multiple ReRAM units, an on-chip cache, a nonlinear activation processing unit, an analog-to-electricity converter, and other peripheral circuits (only the cross array, on-chip cache, and nonlinear activation processing unit are shown in the figure). The DDPG agent is composed of a behavior decision module Actor and a judgment module Critic; the whole pruning and mapping framework includes two levels. In the first level, the behavior decision module Actor of the DDPG agent makes a pruning decision on the neural network model from the first layer to the last layer according to the hardware feed...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides an adjustable hardware perception pruning and mapping framework based on a ReRAM neural network accelerator. The pruning and mapping framework comprises a DDPG agent and the ReRAM neural network accelerator. The DDPG agent is composed of a behavior decision module Actor and an evaluation module Critic, and the behavior decision module Actor is used for making a pruning decision for the neural network; the ReRAM neural network accelerator is used for mapping a model formed under a pruning decision generated by the behavior decision module Actor, and feeding back a performance parameter mapped by the model under the pruning decision to the evaluation module Critic as a signal; the performance parameters comprise energy consumption, delay and model accuracy of the simulator; the judgment module Critic updates a reward function value according to the fed back performance parameters and guides a pruning decision of the behavior decision module Actor in the next stage;according to the method, a pruning scheme which is most matched with hardware and user requirements and is most efficient is made by utilizing a reinforcement learning DDPG agent, so that the delay performance and the energy consumption performance on the hardware are improved while the accuracy is ensured.

Description

technical field [0001] The invention relates to the field of computer science and artificial intelligence, in particular to an adjustable hardware-aware pruning and mapping framework based on a ReRAM neural network accelerator. Background technique [0002] Deep neural networks play an important role in promoting the development of computer vision, natural language processing, robotics and other fields. With the development of mobile Internet of Things platforms, the application of neural networks on IoT devices is developing rapidly. Due to the computation-intensiveness and massive data mobility of neural networks, the application of neural networks will generate high energy consumption and high latency. However, IoT devices require more efficient neural network mapping due to limited computing resources and limited energy support on IoT platforms. solutions to reduce energy consumption and latency. Resistive Random Access Memory (ReRAM) due to its extremely low energy lea...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/08
CPCG06N3/082
Inventor 何水兵杨斯凌陈伟剑陈平陈帅犇银燕龙任祖杰曾令仿杨弢
Owner ZHEJIANG LAB
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products