Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Method, device and equipment for operator scheduling of artificial intelligence model and storage medium

An artificial intelligence and scheduling method technology, applied in the field of data processing, can solve problems such as blocking operator tasks, increasing network inference delay, increasing operator inference delay, etc., to reduce waiting time, avoid inference delay diffusion, and improve AI inference. The effect of efficiency

Active Publication Date: 2022-07-12
SHENZHEN CORERAIN TECH CO LTD
View PDF9 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] At present, most of the strategies adopted for scheduling the operators of the artificial intelligence model put the target operator into the queue according to the number of existing operators in the operator queue. The defect of this scheduling strategy is that when an operator in the operator queue When the running time of an operator is longer than that of other operators in the queue, this operator will cause other operators behind this operator in the queue to have a longer waiting time, thus increasing the number of all subsequent operators. reasoning delay, and because multiple operators of the deep learning neural network in an APP may be distributed to multiple different queues, there may be dependencies between multiple operators in a network. If synchronization is also performed, an operator that takes a long time to complete will not only block operator tasks in this queue, but may also block operator tasks in other queues, resulting in increased inference delays for most networks

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method, device and equipment for operator scheduling of artificial intelligence model and storage medium
  • Method, device and equipment for operator scheduling of artificial intelligence model and storage medium
  • Method, device and equipment for operator scheduling of artificial intelligence model and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] In order to make the purpose, technical solutions and advantages of the present application more clearly understood, the present application will be described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.

[0042] The present application provides an operator scheduling method for an artificial intelligence model. refer to figure 1 As shown, it is a schematic flowchart of a method of an embodiment of an operator scheduling method of an artificial intelligence model of the present application. The method may be performed by an electronic device, which may be imp...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to an operator scheduling method and device of an artificial intelligence model, equipment and a storage medium. The method comprises the following steps: acquiring a reference running time length of each target operator corresponding to a model, calculating an expected queuing time of each operator queue based on the number of operators in a plurality of operator queues and an expected waiting time length of the operators, and calculating the expected queuing time of each operator queue based on the reference running time length and the expected queuing time of each operator queue. And adding each target operator to the operator queue with the shortest expected queuing time. According to the method, the operators in the operator queues can be balanced to the maximum extent, so that the reasoning load on the reasoning engines is balanced to the maximum extent, the model reasoning delay corresponding to the APPs tends to be balanced, the problem of reasoning delay diffusion caused by scheduling according to the number of the operators in the operator queues is effectively avoided, and the reasoning efficiency is improved. The waiting time of each target operator is reduced, and the AI reasoning efficiency of the whole system is improved.

Description

technical field [0001] The present application relates to the technical field of data processing, and in particular, to an operator scheduling method, apparatus, device and storage medium for an artificial intelligence model. Background technique [0002] At present, most of the strategies adopted for scheduling operators of artificial intelligence models put the target operator in the queue according to the number of existing operators in the operator queue. The disadvantage of this scheduling strategy is that when an operator in the operator queue When the running time of the operator is longer than the running time of other operators in the queue, the operator will cause the other operators behind the operator in the queue to have a longer waiting time, thus increasing all subsequent operators. The reasoning delay is high, and since multiple operators of a deep learning neural network in an APP may be distributed to multiple different queues, there may be dependencies bet...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F9/48G06F9/50G06N5/04
CPCG06F9/4881G06F9/4806G06F9/5083G06N5/04
Inventor 伍永情蔡权雄牛昕宇
Owner SHENZHEN CORERAIN TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products