Unlock instant, AI-driven research and patent intelligence for your innovation.

Model reasoning exception handling method and device

An exception handling and model technology, applied in the field of computer networks, can solve problems such as poor applicability, inability to know the task flow on the application side, waste of accelerator computing resources, etc.

Active Publication Date: 2022-06-28
HUAWEI TECH CO LTD
View PDF16 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] The inventor of the present application found in the course of research and practice that in the prior art, when an error occurs in the execution of a certain task flow by the accelerator, the error is only recorded in the task flow, and subsequent tasks will still be executed until the task flow It will greatly waste the computing resources of the accelerator, and may also cause other unpredictable errors due to the execution of wrong tasks by the accelerator.
In addition, since the application side only synchronizes with one task flow in the accelerator, when an error occurs in another task flow, the accelerator cannot return the error information to the application side, and the application side cannot know the task flow where the error occurred, which has poor applicability

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Model reasoning exception handling method and device
  • Model reasoning exception handling method and device
  • Model reasoning exception handling method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0037] The model inference exception processing method and device provided in this application are applicable to all technical fields of inference on models. For the convenience of expression, this application only takes the abnormal processing of AI model inference as an example for illustration. In many application scenarios, such as autonomous driving scenarios, processes such as driver monitoring, parking, and autonomous driving require AI processing of camera images, that is, using AI models for inference. see figure 1 , figure 1 It is a schematic diagram of the application scenario of the AI ​​model inference provided in this application. exist figure 1 The model inference scenario shown may include the cloud server 2000 and the user terminal cluster; the user terminal cluster may include multiple user terminals, such as figure 1 As shown, it specifically includes user terminal 3000a, user terminal 3000b, . . . , user terminal 3000n; figure 1 As shown, the user term...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present application provides a method for handling model inference exceptions, which includes receiving and executing each model inference task of the target model inference session issued by the application processor by the accelerator. The above target model inference session contains multiple target task flows, and one target The task flow contains multiple model inference tasks; when any model inference task in any target task flow in the above-mentioned target model inference session is executed abnormally, the above-mentioned accelerator executes the remaining models contained in the above-mentioned target model inference session according to the exception handling mode Inference tasks, wherein the remaining model inference tasks are all model inference tasks executed after any one of the model inference tasks; the accelerator feeds back the abnormality of the target model inference session to the application processor. By adopting the application, the waste of computing resources caused by model reasoning exceptions can be reduced, the model reasoning abnormalities can be fed back in time, the efficiency of model reasoning can be improved, and the applicability is high.

Description

technical field [0001] The present application relates to the technical field of computer networks, and in particular, to a method and device for abnormal processing of model inference. Background technique [0002] With the development of computer network technology, there are more and more application scenarios that need to build neural network models. For example, in the autonomous driving system, there are a large number of scenarios that need to use artificial intelligence (AI) model inference, and the AI ​​model is essentially a deep neural network model. The neural network model has the characteristics of intensive matrix and vector calculation. The computing power of the system is very high. Ordinary central processing units (Central Processing Units, CPUs) generally cannot meet the computing power requirements of neural network models, so special accelerators are needed to perform inference acceleration, such as graphics processing units (Graphics Processing Units,...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F9/38G06F9/50
CPCG06F9/50
Inventor 朱湘毅
Owner HUAWEI TECH CO LTD