Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-machine collaborative air combat planning method and system based on deep reinforcement learning

A reinforcement learning and planning technology, applied in neural learning methods, stochastic CAD, design optimization/simulation, etc., can solve problems such as large amount of calculation, difficulty in solving, inability to meet real-time decision-making, etc., to improve exploration ability and good training effect Effect

Active Publication Date: 2021-05-28
NAT UNIV OF DEFENSE TECH
View PDF5 Cites 51 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0014] The purpose of the present invention is to provide a multi-aircraft cooperative air combat planning method and system based on deep reinforcement learning, so as to solve the problem of multi-aircraft cooperative air combat planning in the prior art in the real air combat scenario where incomplete information is confronted, the opponent's strategy is unknown and changes in real time. When the system makes air combat decisions, the amount of calculation is large, the solution is difficult, and it cannot meet the technical problems of real-time decision-making

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-machine collaborative air combat planning method and system based on deep reinforcement learning
  • Multi-machine collaborative air combat planning method and system based on deep reinforcement learning
  • Multi-machine collaborative air combat planning method and system based on deep reinforcement learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0049] The technical solutions of the present invention will be clearly and completely described below in conjunction with the accompanying drawings. Apparently, the described embodiments are some of the embodiments of the present invention, but not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0050] Such as Figure 4 The specific battlefield situation information map of the experimental scenario environment is shown. In this scenario, the red and blue forces are configured equally, each containing 3 fighter jets and a base that can take off and land aircraft. The scope of the scenario is a rectangular high sea with a length of 1400 kilometers and a width of 1000 kilometers area. The process of the scenario deduction is that the aircraft takes off from the base, escorts the own base, and destroys t...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

According to the multi-aircraft cooperative air combat planning method and system based on deep reinforcement learning provided by the invention, a combat aircraft is regarded as an intelligent agent, a reinforcement learning agent model is constructed, and a network model is trained through a centralized training-distributed execution architecture, so that the defect that the exploratory performance of a network model is not strong due to low action distinction degree among different entities during multi-aircraft cooperation is overcome; and by embedding expert experience in the reward value, the problem that a large amount of expert experience support is needed in the prior art is solved. Through an experience sharing mechanism, all agents share one set of network parameters and experience playback library, and the problem that the strategy of a single intelligent agent is not only dependent on the feedback of the own strategy and the environment, but also influenced by the behaviors and cooperation relationships of other agents is solved. By increasing the sampling probability of the samples with large absolute values of the advantage values, the samples with extremely large or extremely small reward values can influence the training of the neural network, and the convergence speed of the algorithm is accelerated. The exploration capability of the intelligent agent is improved by adding the strategy entropy.

Description

technical field [0001] The invention belongs to the technical field of space cooperative combat, and in particular relates to a multi-machine cooperative air combat planning method and system based on deep reinforcement learning. Background technique [0002] Since the 1990s, the development of information technology has promoted military reform. The traditional combat style in which each platform uses its own sensors and weapon systems to detect, track, and strike targets can no longer meet the needs of digital warfare. Facing the increasingly complex battlefield environment in modern warfare, a single fighter has limited ability to detect, track and attack targets, and cannot independently complete the designated air-to-air combat missions. Therefore, multiple fighters need to cooperate to maximize combat effectiveness. [0003] Multi-aircraft coordinated air combat refers to a way of warfare in which two or more combat aircraft cooperate and cooperate with each other to ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F30/27G06N3/04G06N3/08G06F111/08
CPCG06F30/27G06N3/084G06F2111/08G06N3/045G05D2101/15G05D2109/22G05D2107/34G05D2105/35G05D1/689G05D1/6983
Inventor 冯旸赫程光权施伟黄魁华黄金才刘忠
Owner NAT UNIV OF DEFENSE TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products