Unlock instant, AI-driven research and patent intelligence for your innovation.

Throughput-optimized deployment method for machine learning inference tasks based on cloud-edge collaboration

A machine learning and throughput technology, applied in instruments, multi-programming devices, data exchange networks, etc., can solve problems such as surge in mobile data traffic, heterogeneous computing equipment resources, computing delay throughput bottlenecks, etc., to achieve optimal throughput volume effect

Active Publication Date: 2021-10-15
JIANGSU ELECTRIC POWER INFORMATION TECH
View PDF8 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the current machine learning system running on terminal devices faces the following three problems: First, the connection of smart devices to the cloud and the surge in mobile data traffic. According to Cisco's forecast, more than 12 billion smart devices will be connected to the network by 2022 Serve
However, the naive model division method may lead to two problems: first, the amount of data transmitted between different stages of the model is too large, which may lead to significant data transmission delay in the hierarchical mobile network environment; second, the cloud edge The computing device resources in the collaborative system are heterogeneous. If the stage with a large amount of computation is deployed on a device with a low computing capacity, the long computation delay of this stage will become the throughput bottleneck of the parallel processing of the pipeline.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Throughput-optimized deployment method for machine learning inference tasks based on cloud-edge collaboration
  • Throughput-optimized deployment method for machine learning inference tasks based on cloud-edge collaboration
  • Throughput-optimized deployment method for machine learning inference tasks based on cloud-edge collaboration

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0034] The present invention will be described in further detail below in conjunction with the accompanying drawings and embodiments, but it should be understood that the description of the following specific embodiments is only to enable those skilled in the art to understand the technical solutions more clearly, rather than to limit the present invention.

[0035] figure 1A deployment architecture diagram of a machine learning inference task in a cloud-edge collaboration scenario provided by an embodiment is shown, and the architecture diagram includes a mobile terminal, a base station, an edge server, and a cloud data center server. The deployment process of machine learning inference tasks can be simplified as follows: the mobile terminal sends the intelligent service request to the nearby base station, and the base station determines the optimal model deployment plan according to the model division strategy, and adopts the pipeline parallel method to execute the inference ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a throughput-optimized machine learning inference task deployment method based on cloud-edge collaboration. Transmission delay, and then establish a throughput-optimized inference task deployment optimization problem, and design an efficient deployment strategy based on the idea of ​​dynamic programming to obtain the optimal deployment plan. Based on the cloud-edge collaboration scenario, the present invention divides the inference task into serial stages, and uses pipelines to process the inference task in parallel, so as to optimize the throughput of the inference task.

Description

technical field [0001] The invention relates to the field of distributed computing and task scheduling, in particular to a machine learning inference task deployment method based on throughput optimization of cloud-edge collaboration. Background technique [0002] With the rapid development of smart devices, the demand for smart services has surged, and more and more smart devices are connected to the network to process the massively generated streaming sensor data. However, the current machine learning system running on terminal devices faces the following three problems: First, the connection of smart devices to the cloud and the surge in mobile data traffic. According to Cisco's forecast, more than 12 billion smart devices will be connected to the network by 2022 Serve. Therefore, sending raw sensor data, such as video from a surveillance camera, to a remote cloud may congest the backhaul network, resulting in reduced throughput and long response times, as well as the ri...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): H04L12/24H04L29/08H04W24/02H04W24/06G06N3/063G06N3/04G06F9/50
CPCH04L41/0823H04L41/083H04L41/145H04L67/10H04W24/02H04W24/06G06F9/5072G06N3/063G06N3/045
Inventor 吴鹏李辉杨定坤
Owner JIANGSU ELECTRIC POWER INFORMATION TECH
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More