The invention discloses a collision avoidance
planning method for mobile robots based on deep
reinforcement learning in a dynamic environment, and belongs to the technical field of
mobile robot navigation. The method of the invention includes the following steps of: collecting
raw data through a
laser rangefinder,
processing the
raw data as input of a neural network, and building an LSTM neural network; through an A3C
algorithm, outputting corresponding parameters by the neural network, and
processing the corresponding parameters to obtain the action of each step of the
robot. The scheme of the invention does not need to model the environment, is more suitable for an unknown obstacle environment, adopts an actor-critic framework and a temporal difference
algorithm, is more suitable for a continuous
motion space while realizing low variance, and realizes the effect of learning while training. The scheme of the invention designs the continuous
motion space with a heading angle limitationand uses 4 threads for
parallel learning and training, so that compared with general deep
reinforcement learning methods, the learning and
training time is greatly improved, the sample correlation isreduced, the high utilization of exploration spaces and the diversity of exploration strategies are guaranteed, and thus the
algorithm convergence, stability and the success rate of
obstacle avoidance can be improved.