Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Mobile robot path planning method based on deep reinforcement learning

A mobile robot and reinforcement learning technology, applied in the direction of instruments, non-electric variable control, two-dimensional position/channel control, etc., can solve the problems of limited generalization ability, poor navigation effect, and low learning efficiency in unfamiliar scenes. Achieve the effects of self-learning efficiency and motion safety improvement, short time spent, and strong action robustness

Active Publication Date: 2021-06-04
CHANGSHA UNIVERSITY OF SCIENCE AND TECHNOLOGY
View PDF8 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] The traditional path planning method in the pedestrian environment is to use a mathematical model or a physical model to construct the interaction state between the robot and pedestrians, and then combine traditional search algorithms such as genetic algorithms to complete the path planning task. Parameters, the generalization ability for unfamiliar scenes is limited, and the navigation effect is not good; with the development of machine learning, data-driven methods have become a popular research direction for robot path planning in pedestrian environments. This method enables robots to have "learning ability". Dadi improves scene adaptability, but also faces problems such as low learning efficiency and difficulty in convergence

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Mobile robot path planning method based on deep reinforcement learning
  • Mobile robot path planning method based on deep reinforcement learning
  • Mobile robot path planning method based on deep reinforcement learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0186] The invention utilizes the deep reinforcement learning algorithm and the artificial potential field method to improve the autonomous learning efficiency of the mobile robot, and can obtain higher action safety and action robustness under training, and the time taken to reach the target position is shorter; the present invention Take the simulation training process on the V-REP software and the testing process of the 3WD omnidirectional wheel mobile robot as examples to elaborate;

[0187] The task scenario designed in this embodiment is that the mobile robot starts from the starting position, passes through five randomly moving pedestrians, and arrives at the target position without collision;

[0188] The method for path planning of a mobile robot in a pedestrian environment based on deep reinforcement learning and artificial potential field method described in this embodiment includes the following steps:

[0189] Step S1. Determine state information according to the ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a mobile robot path planning method based on deep reinforcement learning. The method comprises the steps of S1 determining state information according to a motion scene of a mobile robot; S2 initializing basic parameters of deep reinforcement learning, and pre-training a state value network weight through imitation learning; S3 carrying out forward transmission on the state information through a state value network, and using an epsilon-greedy strategy for guiding the robot to act; S4 obtaining rewards through a comprehensive reward function; S5 continuously updating the weight through the target value network, and updating related parameters; and S6 recording related data and a model which is finally trained in the training process to obtain an optimal navigation strategy of the robot. The method has a path planning scene aiming at the pedestrian environment in the public service field; and a state value network is designed by utilizing an artificial potential field method and an attention mechanism, so that the state information of the robot and a pedestrian is effectively expanded, and the state information interaction is promoted.

Description

technical field [0001] The present invention relates to a mobile robot path planning method in pedestrian environment based on deep reinforcement learning and artificial potential field method, combining data drive and physical model, using artificial potential field method and attention mechanism to design state value network, effectively expanding the The status information of robots and pedestrians promotes the interaction of status information; constructs a new potential field reward function based on the artificial potential field theory, fully considers the position and direction of the robot, and sets rewards in different spaces to balance each part, improving the reward feedback mechanism ; The robot's self-learning efficiency and motion safety are improved, and the time it takes to reach the target position is shorter, and the robot's action is more robust. Background technique [0002] With the rapid development of mobile robot technology, its application scenarios...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G05D1/02
CPCG05D1/0223G05D1/0214G05D1/0221G05D1/0276
Inventor 陈满赖志强李茂军李宜伟李俊日
Owner CHANGSHA UNIVERSITY OF SCIENCE AND TECHNOLOGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products