Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Vehicle reinforcement learning motion planning method based on driving risk analysis

A motion planning and reinforcement learning technology, applied in two-dimensional position/channel control and other directions, can solve the problems of expensive trial and error training, difficult model convergence, lack of interpretability of model output actions, etc., to reduce the incidence of dangerous actions , the model has strong self-learning ability, and the effect of improving generalization ability and interpretability

Pending Publication Date: 2022-07-22
BEIHANG UNIV
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

At present, the research on motion planning based on deep reinforcement learning mostly adopts end-to-end training strategies. The black-box characteristics of the neural network make the model output action lack of interpretability, and the model mostly uses sparse rewards for strategy optimization, making it difficult for the model to converge and requires a lot of effort. A lot of time for trial and error training

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Vehicle reinforcement learning motion planning method based on driving risk analysis
  • Vehicle reinforcement learning motion planning method based on driving risk analysis
  • Vehicle reinforcement learning motion planning method based on driving risk analysis

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0064] Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited by the embodiments set forth herein. Rather, these embodiments are provided so that the present disclosure will be more thoroughly understood, and will fully convey the scope of the present disclosure to those skilled in the art. It should be noted that, unless otherwise specified, the technical or scientific terms used in the present invention should have the usual meanings understood by those skilled in the art to which the present invention belongs.

[0065] figure 1 A flow chart of a vehicle reinforcement learning motion planning method based on driving risk analysis provided according to an embodiment of the present invention. figure 2 The flow...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a vehicle reinforcement learning motion planning method based on driving risk analysis. The method comprises the following steps: acquiring boundary information on two sides of a lane where a vehicle is located, wherein the boundary information consists of horizontal and vertical coordinates (Xr, Xl) of boundary points on two sides of the lane and vertical distances (dl, dr) from the vehicle to boundaries on the left and right sides; a high-precision map and a radar are used for obtaining a state information array which is composed of the coordinate position (X, Y) of a vehicle, the coordinate position (Xobs, Yobs) of an obstacle and the relative speed delta v; the distance delta v is equal to [delta v1, delta v2,..., delta vn], and n is the number of detected obstacles; the sum is combined into state vector relative state information, the state vector relative state information is input into a vehicle motion planning model phi based on deep reinforcement learning, so that a vehicle motion vector action = [a, theta] is output, a belongs to [-amax, amax] is an acceleration action, amax is the maximum acceleration, theta belongs to [-theta max, theta max] is a steering wheel turning action, when theta belongs to [0, theta max], theta is turned leftwards, and when theta belongs to [-theta max, 0], theta is turned rightwards; after the vehicle runs for t duration according to the obtained action, whether the vehicle arrives at the destination or not is judged, if yes, the working state is ended, and if not, the step 1 is executed for vehicle control of the next time step length.

Description

technical field [0001] The invention belongs to the field of unmanned vehicle motion planning, and more particularly relates to a vehicle reinforcement learning motion planning method based on driving risk analysis. Background technique [0002] The motion planning module is an important technical link of unmanned driving. In the unmanned driving technology chain, it is connected to the environment perception module and the execution control module. It is the intelligent command center of the unmanned vehicle. In the past, the research on unmanned motion planning mainly used rule-based methods, supervised learning methods and optimization methods. The rule-based method can only model parameters for a simple specific driving environment based on certain assumptions, such as following a car and changing lanes. , obstacle avoidance and other specific scenarios, but it is not suitable for complex and changeable urban driving scenarios; the supervised learning method needs to col...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G05D1/02
CPCG05D1/0214G05D1/0221
Inventor 周彬廖亚萍余贵珍倪浩原张传莹
Owner BEIHANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products