Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Autonomous equipment decision control method based on distributed reinforcement learning

A technology of autonomous equipment and reinforcement learning, applied in the direction of adaptive control, comprehensive factory control, general control system, etc., can solve problems such as difficult to achieve results, limited ability of deep reinforcement learning model, and achieve the effect of avoiding slow training speed

Pending Publication Date: 2022-08-02
NANJING UNIV
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

With the trend of comprehensive and complex research on decision-making control of autonomous equipment, it is difficult to achieve results in a limited time for autonomous equipment decision-making control agent training using only a stand-alone method.
[0003] In existing reinforcement learning solutions, the training of autonomous device decision-making control agents is severely constrained by limited computing resources. When faced with more complex and more realistic problem scenarios, the capabilities of deep reinforcement learning models that can be trained are limited. Often only solve single-field problems such as obstacle avoidance, path planning, and dynamic control

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Autonomous equipment decision control method based on distributed reinforcement learning
  • Autonomous equipment decision control method based on distributed reinforcement learning
  • Autonomous equipment decision control method based on distributed reinforcement learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0043] Below in conjunction with specific embodiments, the present invention will be further illustrated, and it should be understood that these embodiments are only used to illustrate the present invention and not to limit the scope of the present invention. The modifications all fall within the scope defined by the appended claims of this application.

[0044] A decision-making control method for autonomous equipment based on distributed reinforcement learning. First, a training system including sampling nodes, cache nodes and training nodes is established, and then distributed asynchronous maximum entropy training is performed. Finally, the training results are compiled into efficient concurrent autonomous equipment decision-making. control module. Including training system building steps, distributed training steps and concurrent acceleration model export steps

[0045] The steps to build the training system are as follows: figure 1shown. In order to be able to run mult...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an autonomous equipment decision control method based on distributed reinforcement learning. The method comprises a training environment building step, a distributed training step and a decision model outputting step. A sampling node, a cache node and a training node are configured in a distributed cluster, an autonomous equipment simulation environment is packaged into a container mirror image, and virtual resources are allocated. A distributed agent training system is started, training end model parameters are initialized, the parameters are sent to a cache node and then forwarded to a sampling node, after the interaction process of an agent and a simulation environment is completed, data are returned to the training node, and the model parameters are updated by using a deorbit deep reinforcement learning algorithm. And after the distributed training is completed, exporting the model from the system, and switching to a rapid reasoning mode for intelligent decision-making. According to the method, training can be carried out on a large-scale distributed cluster in the implementation process, hardware resources can be fully utilized, and meanwhile the bandwidth requirement can be remarkably reduced for improvement of a communication mode.

Description

technical field [0001] The invention relates to a decision-making control method for autonomous equipment based on distributed reinforcement learning, and belongs to the technical field of autonomous equipment control and distributed systems. Background technique [0002] In reality, the realization of autonomous equipment decision control based on reinforcement learning has the problem of large data demand. With the trend of integration and complexity of research on decision-making control of autonomous equipment, it is difficult to achieve results in a limited time by only using a single-machine method for intelligent agent training of decision-making control of autonomous equipment. [0003] In the existing reinforcement learning solutions, the training of autonomous equipment decision control agents is severely restricted by limited computing resources. When faced with more complex and more real problem scenarios, the ability of deep reinforcement learning models that ca...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G05B13/04
CPCG05B13/042Y02P90/02
Inventor 詹德川张云天俞扬周志华
Owner NANJING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products