Eureka AIR delivers breakthrough ideas for toughest innovation challenges, trusted by R&D personnel around the world.

Quantification and hardware acceleration method and device for multi-task neural network

A neural network and hardware acceleration technology, applied in the field of multi-task neural network quantization and hardware acceleration, can solve problems such as large hardware overhead

Pending Publication Date: 2020-10-20
宁波物栖科技有限公司
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] During the training process of the neural network model, 32bit floating-point numbers (fp32) are usually used to save the parameters and activation values ​​of the model, but the storage, transmission, and operation of floating-point numbers will bring greater hardware overhead than fixed-point numbers.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Quantification and hardware acceleration method and device for multi-task neural network
  • Quantification and hardware acceleration method and device for multi-task neural network
  • Quantification and hardware acceleration method and device for multi-task neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0044]In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments It is a part of embodiments of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0045] In order to facilitate the understanding of the embodiments of the present invention, further explanations will be given below with specific embodiments in conjunction with the accompanying drawings, which are not intended to limit the embodiments of the present invention.

[0046] Such as figure 1 As shown, it is a schematic diagram o...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention relates to a quantification and hardware acceleration method and device for a multi-task neural network. The method comprises steps of obtaining a plurality of preset multi-task neural network models, and enabling the number of layers of the multi-task neural network models to be consistent; extracting a shared weight of the same layer in the plurality of multi-taskneural network models, and generating a corresponding weight code table, the weight code table including the shared weight; training samples of the multiple multi-task neural network models are obtained, shared feature values being analyzed from the training samples, a corresponding feature value code table being generated, and the feature value code table comprising the shared feature values; and putting a plurality of multi-task neural network models into neural network accelerator hardware, and accelerating a plurality of multi-task neural network models based on the weight code table andthe feature value code table.

Description

technical field [0001] Embodiments of the present invention relate to the field of hardware acceleration of deep neural network models, and in particular, to a quantization and hardware acceleration method and device for multi-task neural networks. Background technique [0002] Deep neural networks have achieved good results in tasks such as image processing, speech recognition, and natural language processing. At the same time, they have been widely used in scenarios such as terminal devices and cloud data centers, and there have been many dedicated methods for accelerating deep neural networks. There are hardware designs, namely neural network accelerators. The deep neural network model usually has a huge amount of parameters and calculations, and the multiplication and addition operations occupy the vast majority of the calculations. Therefore, the neural network accelerator usually uses an array multiply and accumulate (MAC, multiply and accumulate) to Improve computing...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063G06N3/04G06F7/544G06F1/03
CPCG06N3/063G06F7/5443G06F1/03G06N3/045
Inventor 翟乃鹏
Owner 宁波物栖科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Eureka Blog
Learn More
PatSnap group products