Unlock instant, AI-driven research and patent intelligence for your innovation.

Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning

A technology of reinforcement learning and control methods, which is applied in the control of finding targets, vehicle position/route/height control, non-electric variable control, etc. Effects of autonomy and efficiency of landing

Active Publication Date: 2020-07-28
BEIHANG UNIV
View PDF7 Cites 23 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

For example, the method disclosed in the document "Carlos Sampedro, Alejandro Rodriguez-Ramos, Ignacio Gil, Luis Mejias, Pascual Campoy. Image-Based Visual Servoing Controller for Multirotor Aerial Robots Using Deep Reinforcement Learning[J].IEEE.2018", which will firstly learn from The image obtained in Gazebo is converted into the state deviation between the UAV and the landing point, and then the control amount required for the movement of the UAV is obtained through the DDPG method, but this method is relatively cumbersome; or, it is directly output through the image information. Continuous direction instructions, such as the literature "Riccardo Polvara, Massimiliano Patacchiola, Sanjay Sharma, Jian Wan, Andrew Manning, Robert Sutton, Angelo Cangelosi. Toward End-to-End Control for UAV Autonomous Landing via Deep Reinforcement Learning[J].IEEE.2018" A deep reinforcement learning based approach to UAV landing control is proposed based on a hierarchical structure of Deep Q-Networks (DQN), which are used as high-end control strategies for navigation in different phases, including front-back, left-right Control commands such as descent, but this method is less accurate

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning
  • Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning
  • Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0047] The traditional autonomous landing algorithm is generally divided into four steps: obtain the required observations in the environment, perform state estimation, modeling and prediction from the observations, and finally perform landing planning control, and the end-to-end deep reinforcement learning algorithm is Using the network to replace the intermediate steps in the traditional autonomous landing, the landing planning control can be directly obtained from the observations, which can greatly simplify the landing process and be more in line with human thinking. figure 1 Shows the difference between traditional landing algorithms and end-to-end deep reinforcement learning algorithms. Based on an end-to-end deep reinforcement learning neural network, the present invention provides an end-to-end control method for a drone.

[0048] The present invention will be further described below in conjunction with the accompanying drawings and examples. It should be understood th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the field of unmanned aerial vehicle autonomous control, and relates to an unmanned aerial vehicle end-to-end control method based on deep reinforcement learning. The invention discloses an unmanned aerial vehicle end-to-end control method based on deep reinforcement learning. The method comprises the steps of designing a deep reinforcement learning neural network; designing an end-to-end reward and punishment function rland for training unmanned aerial vehicle landing trajectory control of the designed deep reinforcement learning neural network; building an unmanned aerial vehicle landing scene in the simulation environment; and training the designed deep reinforcement learning neural network based on the simulation environment with the built unmanned aerial vehicle landing scene. According to the invention, the processed image obtained by the airborne camera of the unmanned aerial vehicle is used as input, and the unmanned aerial vehicle control instruction can be directly obtained after deep reinforcement learning network processing, so that autonomous landing of the unmanned aerial vehicle is realized.

Description

technical field [0001] The invention belongs to the field of autonomous control of unmanned aerial vehicles, and relates to an end-to-end control method for unmanned aerial vehicles based on deep reinforcement learning. Background technique [0002] The current problem of UAV autonomous landing control mainly uses hand-crafted geometric features and sensor data fusion to identify fiducial markers and guide the UAV to move in its direction. For example, the method disclosed in the document "Carlos Sampedro, Alejandro Rodriguez-Ramos, Ignacio Gil, Luis Mejias, Pascual Campoy. Image-Based Visual Servoing Controller for Multirotor Aerial Robots Using Deep Reinforcement Learning[J].IEEE.2018", which will firstly learn from The image obtained in Gazebo is converted into the state deviation between the UAV and the landing point, and then the control amount required for the movement of the UAV is obtained through the DDPG method, but this method is cumbersome; Continuous direction ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F30/20G06N3/04G05D1/10G05D1/12
CPCG05D1/101G05D1/12G06N3/045
Inventor 赵江王隆洪蔡志浩王英勋
Owner BEIHANG UNIV