Multi-unmanned aerial vehicle local route planning method and device based on improved Q-Learning
A multi-UAV, route planning technology, applied in neural learning methods, based on specific mathematical models, navigation calculation tools, etc., can solve the problems of long planning time, low algorithm convergence speed, long path search time, etc., to achieve rapid planning Effect
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0065] A kind of multi-UAV local route planning method based on improved Q-Learning of the present invention, see figure 1 shown, including
[0066] Step S1, obtaining the location of the newly added sudden threat source, determining the newly added sudden threat area in the flight environment, and then determining the starting point and target point of the local route replanning that can avoid the newly added sudden threat area.
[0067] The specific implementation process is:
[0068] According to the nature of multi-UAV collaborative local route planning tasks, when multi-UAVs fly to a certain waypoint along the global route, they use the sensor to detect the surrounding environment to collect information such as the position coordinates and radius of the sudden threat source. The local flight environment of the aircraft is updated.
[0069] The process of updating the UAV flight environment is to read the new mission environment. In the two-dimensional environment space...
Embodiment 2
[0128] The embodiment of the present invention is simulation analysis:
[0129] Figure 4 It is a comparison result diagram of the convergence experiment of the Q-Learning algorithm and the Q-Learning algorithm based on the CMAC neural network in the same simulation environment in the specific embodiment of the present invention. The horizontal axis in the figure is the number of iterations, that is, the number of algorithm learning times, and the vertical axis is Cumulative remuneration value. Depend on Figure 4 It can be seen that both algorithms converge, but the convergence speeds of the algorithms are different. Based on the Q-Learning algorithm, when the number of learning times is 120, the cumulative reward reaches the maximum value and no longer increases. At this time, the cumulative reward value is about 62; when the Q-Learning algorithm based on the CMAC neural network is 50 times, the cumulative reward The reward has reached the maximum value, and the cumulativ...
Embodiment 3
[0146] Correspondingly, an improved Q-Learning-based multi-UAV local route planning device of the present invention includes:
[0147] According to the location of the newly added sudden threat source, determine the newly added sudden threat area in the flight environment, and then determine the starting point and target point of the local route replanning that can avoid the newly added sudden threat area;
[0148] According to the starting point, target point and threat area of multi-UAV local route re-planning, the multi-UAV local route planning system model in the process of multi-UAV local re-planning movement is determined;
[0149] Using the improved Q-Learning algorithm of the generalization ability of the cerebellar neural network to solve the above multi-UAV local route planning system model, and plan the optimal local route of multi-UAV;
[0150] According to the length of the re-planned local route section of the multi-UAV, adjust the speed of the multi-UAV when f...
PUM
Abstract
Description
Claims
Application Information
- R&D Engineer
- R&D Manager
- IP Professional
- Industry Leading Data Capabilities
- Powerful AI technology
- Patent DNA Extraction
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic, Popular Technical Reports.
© 2024 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap|About US| Contact US: help@patsnap.com