Unmanned aerial vehicle path planning method based on transfer learning strategy deep Q-network

A technology of transfer learning and path planning, applied in two-dimensional position/channel control, vehicle position/route/altitude control, non-electric variable control, etc. The effect of improving the speed of convergence, shortening the time spent

Active Publication Date: 2020-01-17
NANJING UNIV OF AERONAUTICS & ASTRONAUTICS
View PDF8 Cites 20 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to provide a UAV path planning method combining transfer learning and DQN algorithm, which can solve the problems of slow convergence speed and low success rate when DQN algorithm performs path planning in a dynamic environment

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Unmanned aerial vehicle path planning method based on transfer learning strategy deep Q-network
  • Unmanned aerial vehicle path planning method based on transfer learning strategy deep Q-network
  • Unmanned aerial vehicle path planning method based on transfer learning strategy deep Q-network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0026] The technical solutions of the present invention are described in detail with reference to the accompanying drawings.

[0027] A UAV path planning method based on the migration learning strategy deep Q network of the present invention specifically includes the following steps:

[0028] Step 1, use the grid method to model and describe the dynamic environment in which the UAV is located.

[0029] (1.1) The dynamic environment in which the UAV is located is a 20x20 grid map, such as figure 2 shown. Among them, the light pink squares are movable obstacles; the other black positions are immovable obstacles, which are L-shaped wall, horizontal wall, vertical wall, T-line wall, inclined wall, square wall and irregular wall. Test the obstacle avoidance effect of the agent; the yellow circle is the target position, the red square is the starting position of the agent, the target position and the starting position of the agent can be randomly generated, when the agent moves t...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses an unmanned aerial vehicle path planning method based on a transfer learning strategy deep Q-network (DQN). The method comprises the steps of: firstly carrying out the modelingand description of a dynamic environment where a UAV is located through adoption of a grid method, and building a state space model and an action space model of the UAV; secondly, initializing network parameters of a DQN and the current state of the unmanned aerial vehicle; training the DQN by adopting a return mechanism based on a social force model under the static environment model to obtain anetwork weight and an optimal action value; migrating the network weight and the optimal action value obtained by training in the static environment to the dynamic environment by using transfer learning, and continuing neural network training to obtain an action to be executed by the UAV; and finally, calculating the position of the unmanned aerial vehicle at the current moment to achieve path planning of the unmanned aerial vehicle in the dynamic environment. According to the method, the problems of the low DQN training convergence rate, the non-ideal path planning and the low success rate when the unmanned aerial vehicle performs path planning in a dynamic environment are effectively solved.

Description

technical field [0001] The invention belongs to the field of UAV path planning, and in particular relates to a UAV path planning method based on migration learning and DQN (Deep Q-Network), applying migration learning and deep reinforcement learning to carry out UAV path planning in a dynamic environment . [0002] technical background [0003] UAV path planning is a core issue in the field of UAV technology research, and related algorithms are developing rapidly. Traditional methods include: Dijkstra's shortest path search method (greedy algorithm), A* algorithm, ant colony optimization algorithm, reinforcement learning algorithm, etc. The core idea of ​​Dijkstra's algorithm is that the next vertex selected by each exploration is the point closest to the Euclidean distance from the starting point until the target is found. This method is only suitable for static maps with known overall information, and the efficiency is low; the A* algorithm is based on the Dijkstra method...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G05D1/02
CPCG05D1/0223G05D1/0214G05D1/0221
Inventor 丁勇汪常建胡佩瑶
Owner NANJING UNIV OF AERONAUTICS & ASTRONAUTICS
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products