Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Piloted driving vehicle training method based on virtual environment and depth double-Q network

A virtual environment and automatic driving technology, applied in the direction of probability network, neural learning method, based on specific mathematical models, etc., can solve the problems of poor robustness, labor-intensive data collection, and variation, etc., to achieve strong robustness, The training process is fast and stable, avoiding a large amount of workload and the effect of high requirements for human manipulation

Pending Publication Date: 2020-02-28
NORTH CHINA UNIVERSITY OF TECHNOLOGY
View PDF6 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The inventive method needs artificial collection of training data according to the current feature model and environmental data, which requires a lot of artificial data collection work
The car trained by manually collected data has poor robustness, and the performance of the model will deteriorate after a slight change in the environment

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Piloted driving vehicle training method based on virtual environment and depth double-Q network
  • Piloted driving vehicle training method based on virtual environment and depth double-Q network
  • Piloted driving vehicle training method based on virtual environment and depth double-Q network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0039] The specific implementation manner of the present invention will be described below with the algorithm steps in conjunction with the accompanying drawings, so that those skilled in the art can better understand the present invention.

[0040] The present invention proposes a self-driving car training method based on a virtual environment and a deep double-Q network, such as figure 1 shown in the following steps:

[0041] Step 1: Refer to the real track, preset the environmental parameters, and build a virtual environment for the car track suitable for reinforcement learning training based on Unity;

[0042] In this embodiment, run under the Linux system, download and configure Unity and OpenAI gym. Using the game engine sandbox in Unity, according to the size of the physical car, set the road property to a two-way street with a width of 60cm. The size of the car in the virtual environment is set to a ratio of 1:16. The frame skipping parameter in the Unity environment ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a piloted driving vehicle training method based on a virtual environment and a depth double-Q network. The method comprises: constructing a vehicle racing track virtual environment based on Unity; establishing a communication connection between the Unity and a piloted driving vehicle model under a Websocket protocol; establishing a Markov model that represents a vehicle piloted driving process, transforming a depth double-Q network algorithm to train a piloted driving model; configuring an entity vehicle soft hardware environment, migrating a trained vehicle piloted driving model; and testing an entity vehicle in a real racing track to train a piloted driving operation of the model. According to the method, a virtual environment is used for training a model, to implement robustness of a piloted driving training algorithm model on a complex environment and road condition. According to the method, by means of the depth double-Q network algorithm, a problem of high train complexity caused by a large assessment value of a Q value is avoided, so that a training process is simple and fast. The method has advantages of strong robustness, a fast speed, low costs and the like. The method can be applied to training and learning of unpiloted operation in fields such as intelligent traffic, aerospace, robots and the like.

Description

technical field [0001] The invention relates to a self-driving car training method based on a virtual environment and a deep double-Q network, belongs to the technical field of automation control, and specifically relates to a deep reinforcement learning algorithm. Background technique [0002] In recent years, with the rapid development of artificial intelligence technology, especially deep learning technology, major breakthroughs have been made in the fields of target recognition and intelligent control. Deep learning technology, such as convolutional neural network (CNN), is widely used in various types of vision-related problems, including the field of automatic driving, and the use of deep learning technology to achieve automatic driving is the mainstream of current research on automatic driving technology. At the same time, cars have entered thousands of households. With the increase of car ownership, the incidence of traffic safety accidents is also gradually increasi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G05D1/02G06N3/04G06N3/08G06N7/00G06T7/12G06T7/13G06T7/168G06T17/05
CPCG05D1/0221G06T17/05G06N3/08G06T7/12G06T7/13G06T7/168G06T2207/20061G06N7/01G06N3/045
Inventor 杜涛张琪田常正
Owner NORTH CHINA UNIVERSITY OF TECHNOLOGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products