Unlock instant, AI-driven research and patent intelligence for your innovation.

Training method, device and chip of a neural network model

A technology of neural network model and training method, which is applied to the training method of neural network model, device and chip field, can solve the problem of large information communication volume, and achieve the effect of reducing communication volume

Active Publication Date: 2021-01-12
HUAWEI TECH CO LTD
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In the above scheme, since the L layer of the neural network model includes a large number of model parameters, the application of this scheme will cause each working module to push a large number of local gradients of model parameters to the server module, and pull down a large number of model parameters from the server module. Global gradient, resulting in a large amount of information communication between the server module and each working module

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Training method, device and chip of a neural network model
  • Training method, device and chip of a neural network model
  • Training method, device and chip of a neural network model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0045] In order to make the object, technical solution and beneficial effects of the present invention more clear, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0046] figure 2 It exemplarily shows a schematic diagram of an application scenario architecture applicable to the embodiment of the present invention, as shown in figure 2 As shown, there will be a variety of raw data in the specific implementation, such as figure 2The telecom data 201, financial data 202, consumer data 203, etc. in the big data platform 204 perform data collection, data storage and data calculation on these raw data, and obtain the data processed by the big data platform 204. The data mining platform 205 acquires data processed by the big data platform 204 from th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiments of the present invention relate to the field of machine learning, and in particular to a neural network model training method, device and chip, which are used to reduce the communication traffic between the server module and each working module during the training process of the neural network model. In the embodiment of the present invention, the model training method of each layer is determined according to the estimated data amount in the model parameter set of each layer and the estimated data amount of the output data. When the jth layer is a model parallel training method, due to the The second output data is the output data of the j-1 layer training of the m work modules, so the work module performs model parameter training according to the second output data, and can directly obtain the global gradient of the model parameters, compared to the work module in the prior art. The server module pushes up the local gradient of the model parameters and pulls down the global gradient of the model parameters from the server module to obtain the global gradient of the model parameters, which reduces the amount of communication between the working module and the server module.

Description

technical field [0001] Embodiments of the present invention relate to the field of neural network model training, and in particular to a neural network model training method, device, and chip. Background technique [0002] Since deep learning has achieved great success in large-scale image classification datasets, both academia, government and industry are vigorously promoting the development of deep learning, and have continuously achieved new results. As the main model form in deep learning, the feedforward neural network model has been widely used in tasks such as face recognition, image classification, target detection, and video analysis, and is being rapidly adopted by major machine vision manufacturers for intelligent Image, video processing and other products. At present, the depth of the feed-forward neural network model is getting deeper and deeper, and the structure is more and more complex. For example, in many intelligent image and video processing tasks, the d...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/04
CPCG06N3/04G06N3/063G06N3/084G06N3/044G06N3/045
Inventor 白小龙张长征夏命榛
Owner HUAWEI TECH CO LTD