Vehicle-mounted edge task centralized scheduling and resource allocation joint optimization method based on deep reinforcement learning
A technology of reinforcement learning and resource allocation, applied in the field of in-vehicle mobile edge computing, which can solve problems such as server load imbalance
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment approach
[0082] Such as figure 1 As shown, it is assumed that vehicle j will carry task Q at this time j Send to RSU, then according to the specific implementation mode of the present invention is as follows:
[0083] (1) Use the SDN controller to collect relevant information. The set of edge servers in each local area network is ser, the clock cycle set h of the edge server, the CPU occupancy rate set util of the edge server, the vehicle task set q to be processed, and the CPU cycle set m occupied by each vehicle task;
[0084] (2) According to the data obtained in (1), calculate the task Q j The calculation delay of:
[0085]
[0086] (3) SDN summarizes the information of other vehicles and edge servers, and calculates the calculation delay of vehicle tasks in all servers:
[0087]
[0088] (4) SDN summarizes the load information of the edge server, and converts the decision-making method of on-board task edge scheduling and resource allocation into solving the following ma...
specific Embodiment approach
[0090] (5) Use the DDQN algorithm to solve the mathematical problem in (4). The specific implementation is as follows:
[0091] 1. First obtain the initialization state, that is, the current vehicle task and the relevant information of the edge server. The current Q network generates action A according to the state S, and action A is the computing resource allocated to each task. The specific method is A=maxQ(φ(S),a,ω), which means that in the current state S, the neural network ω selects the action with the largest Q value from all actions a according to the feature vector φ(S) of the state S .
[0092] 2. Calculate the reward R according to the state S and action A, and generate a new state S'. After calculating the current on-board tasks, the number of on-board tasks waiting to be calculated and the various states of the edge server have changed, and the new state is S';
[0093] 3. Store the previously obtained {φ(S), A, R, φ(S')} into the experience playback pool, whi...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More - R&D
- Intellectual Property
- Life Sciences
- Materials
- Tech Scout
- Unparalleled Data Quality
- Higher Quality Content
- 60% Fewer Hallucinations
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic, Popular Technical Reports.
© 2025 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap|About US| Contact US: help@patsnap.com



