Power grid real-time adaptive decision-making method based on deep reinforcement learning

A technology of reinforcement learning and decision-making method, applied in the field of deep reinforcement learning, which can solve the problem of new energy consumption without fully considering the robust operation of the new power grid, and cannot solve the problem of explosive action space of the new power grid system optimization strategy exploration vulnerability Problems, no identical or similar problems were found, and good control effects were achieved

Pending Publication Date: 2022-03-22
ELECTRIC POWER SCI & RES INST OF STATE GRID TIANJIN ELECTRIC POWER CO +2
View PDF8 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] (1) The existing technology does not fully consider the robust operation of the new power grid when there is a sudden failure under the random fluctuation of new energy and the consumption of new energy when the proportion of new energy is high;
[0006] (2) Existing technologies cannot solve the problem of explosive action space in the decision-making process of adaptive unit scheduling based on deep reinforcement learning for new power grids and the vulnerability of power grid systems in the process of exploring optimization strategies
[0007] After searching, no documents of the prior art identical or similar to the present invention were found

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Power grid real-time adaptive decision-making method based on deep reinforcement learning
  • Power grid real-time adaptive decision-making method based on deep reinforcement learning
  • Power grid real-time adaptive decision-making method based on deep reinforcement learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0102] Embodiments of the present invention are described in further detail below:

[0103] A real-time adaptive decision-making method for power grids based on deep reinforcement learning, such as figure 1 shown, including the following steps:

[0104] Step 1. Model the adaptive scheduling problem of the new power system unit as a Markov decision process (Markov Decision Process, MDP);

[0105] The concrete steps of described step 1 include:

[0106] Use a 4-dimensional tuple to describe (S, A, P, R), where S represents the state set of the power grid system, A represents the action set of the power grid system, and P: S×A×S→[0,1] represents the state Transition probability, R:S×A→R means reward mechanism;

[0107] In this embodiment, step 1 involves the construction of an MDP model. Many control decision-making problems in the power grid can be described as MDP models, which are used to solve discrete time-sequence control problems in stochastic dynamic environments, spec...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention relates to a power grid real-time adaptive decision-making method based on deep reinforcement learning. The method comprises the following steps: step 1, modeling a novel power system unit adaptive scheduling problem as a Markov decision-making process; 2, researching the basic principle of an SAC algorithm, and solving a strategy enabling the accumulated reward value of the MDP model in the step 1 to be maximum; step 3, designing a neural network pre-training scheme based on behavior cloning in the IL, simulating expert experience, optimizing an original action space, proposing an IL-SAC algorithm, training a corresponding power grid optimization scheduling agent based on the IL-SAC algorithm and 105 real power grid scene data, and performing optimization scheduling on the power grid. During testing, the intelligent agent can output a real-time decision scheme in response to different power grid scene data, and intelligent regulation and control of a novel power grid system are realized. The power grid dispatching strategy can be output in real time.

Description

technical field [0001] The invention belongs to the technical field of deep reinforcement learning, and relates to a real-time self-adaptive decision-making method for a power grid, in particular to a real-time self-adaptive decision-making method for a power grid based on deep reinforcement learning. Background technique [0002] With the continuous development of the social economy and the continuous construction of industrial modernization, the demand for energy continues to grow, and energy problems have gradually emerged. To meet the needs of social development, my country's new energy industry has developed very rapidly. Behind the rapid development of new energy, there are It is the problem of overproduction of new energy. In the development of new energy such as hydropower and wind power, these problems are more serious and prominent. Building a smart grid operation mode that maximizes the consumption of new energy is becoming a complex task beyond human expertise. T...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G05B13/02G05B13/04G06N3/08
CPCG05B13/029G05B13/04G06N3/08Y02E40/70Y04S10/50
Inventor 马世乾陈建商敬安崇志强王天昊韩磊吴彬李昂张志军董佳孙峤郭凌旭黄家凯袁中琛穆朝絮韩枭赟徐娜
Owner ELECTRIC POWER SCI & RES INST OF STATE GRID TIANJIN ELECTRIC POWER CO
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products