Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Neural network accelerator model conversion method and device

A neural network model and neural network technology, applied in the field of neural network accelerator model conversion methods and devices, can solve the problems of cumbersome hardware, large limitations, hindering the deployment efficiency of neural network models, etc., and achieve the effect of reducing reasoning time and number

Pending Publication Date: 2022-01-18
ZHEJIANG LAB +1
View PDF0 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] The current neural network technology is developing rapidly, and various machine learning frameworks have emerged as the times require. Various machine learning frameworks have their own advantages and disadvantages. Therefore, in academia and industry, various machine frameworks coexist, which has also led to differences. The models trained by the framework have different file formats. For ASIC (Application Specific Integrated Circuit, Application Specific Integrated Circuit) devices such as neural network accelerators, various model file formats need to be adapted, which greatly hinders the neural network model from being used in ASIC devices. deployment efficiency
[0003] Existing model format files must be read based on the framework or protocol corresponding to the format. For example, the TensorFlow model file uses Protocol Buffers. If you want to quickly read and obtain the network layer information in the model file, you must use TensorFlow Software packages are required, which is very cumbersome on the hardware; at the same time, the model format saved by most machine learning frameworks contains a lot of information that is useless for the neural network accelerator to run the model (such as the learning rate required for training, batch data size, etc. ), which are useless information in neural network accelerators (mainly used for model inference)
In short, the traditional neural network accelerator must use the machine learning framework program to use its saved model files, and the use of the framework for the neural network accelerator will have the problem of adapting different frameworks, which has great limitations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Neural network accelerator model conversion method and device
  • Neural network accelerator model conversion method and device
  • Neural network accelerator model conversion method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0029] Specific embodiments of the present invention will be described in detail below in conjunction with the accompanying drawings. It should be understood that the specific embodiments described here are only used to illustrate and explain the present invention, and are not intended to limit the present invention.

[0030] The present invention adds an agent between the machine learning framework and the neural network accelerator, for the machine learning framework, for the files saved in several mainstream machine learning frameworks, using software to read their network layer information, and for the neural network The accelerator can save the read network layer information in a data format and supported operator types that are convenient for the neural network accelerator to read the network layer information. In this way, the neural network model parameter files in different formats are converted into file formats suitable for neural network accelerators (including dat...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a neural network accelerator model conversion method and device. The method comprises the steps of obtaining a to-be-converted neural network model; analyzing the model network structure file to obtain all network layers of the model; reconstructing the network layer; mapping to operator nodes supported by a neural network accelerator; and finally, serializing the converted operator nodes and model weights according to a network topology structure to generate a target file. The device comprises a neural network model construction module, a reconstruction module, a mapping module and a serialization module. According to the invention, the problem of multi-adaptation difficulty of deployment of various format models on neural network accelerator equipment is solved, model conversion can be efficiently carried out, and the model format adaptive to the neural network accelerator is generated.

Description

technical field [0001] The invention relates to the field of artificial intelligence, in particular to a neural network accelerator model conversion method and device. Background technique [0002] The current neural network technology is developing rapidly, and various machine learning frameworks have emerged as the times require. Various machine learning frameworks have their own advantages and disadvantages. Therefore, in academia and industry, various machine frameworks coexist, which has also led to differences. The models trained by the framework have different file formats. For ASIC (Application Specific Integrated Circuit, Application Specific Integrated Circuit) devices such as neural network accelerators, various model file formats need to be adapted, which greatly hinders the neural network model from being used in ASIC devices. deployment efficiency. [0003] Existing model format files must be read based on the framework or protocol corresponding to the format....

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/08G06N3/045
Inventor 凡军海朱国权杨方超陆启明金孝飞孙世春章明何煜坤马德胡有能
Owner ZHEJIANG LAB
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products