Unlock instant, AI-driven research and patent intelligence for your innovation.

Production line parallel training node weight distribution method based on version difference

A node weight and pipeline technology, which is applied in the field of pipeline parallel training node weight distribution, can solve the problem of low precision and achieve the effect of improving model precision, ensuring effectiveness and improving accuracy

Inactive Publication Date: 2021-10-01
HOHAI UNIV +1
View PDF5 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Purpose of the invention: In order to solve the problem of outdated node weights in pipeline parallel training and the low accuracy of existing node weight prediction methods, the present invention provides a method for assigning node weights for pipeline parallel training based on version differences, based on asynchronous pipeline parallel training , making the node weight prediction more accurate, further promoting the improvement of the accuracy of the deep learning model, and ensuring the effectiveness of the model training

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Production line parallel training node weight distribution method based on version difference
  • Production line parallel training node weight distribution method based on version difference
  • Production line parallel training node weight distribution method based on version difference

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] Below in conjunction with specific embodiment, further illustrate the present invention, should be understood that these embodiments are only used to illustrate the present invention and are not intended to limit the scope of the present invention, after having read the present invention, those skilled in the art will understand various equivalent forms of the present invention All modifications fall within the scope defined by the appended claims of the present application.

[0060] Aiming at the staleness of node weights in pipeline parallel training and the low accuracy of existing node weight prediction methods, based on asynchronous pipeline parallel training, a more accurate weight prediction method is used to calculate the difference value of weight versions and improve the accuracy of node weight prediction , to further improve the model accuracy, achieve good node weight update, and further ensure the effectiveness of model training.

[0061] figure 1 (a) Differ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a version difference-based assembly line parallel training node weight distribution method. The method comprises the following steps of: loading an initialized deep learning model; constructing a pipeline parallel training scheme; by using an asynchronous parameter updating method, executing the training of different batches concurrently, and recording the training batches to complete forward and backward transmission processes within unit pipeline execution time; and predicting a node weight in a future training process by using the latest training node weight version, and performing version difference calculation of the batch according to node weight prediction; and after the node completes version difference calculation of all batches, completing prediction weight updating of the node. The version difference of all the nodes is calculated, that is, assembly line parallel training node weight distribution is completed; besides, the data is deployed to heterogeneous computing nodes to obtain an assembly line parallel training node weight allocation scheme for the to-be-trained target network. According to the method, the node weight prediction is more accurate.

Description

technical field [0001] The invention relates to a weight distribution method for pipeline parallel training nodes based on version differences, and belongs to the technical field of computer pipeline system optimization. Background technique [0002] Deep neural networks are widely used in various fields, and have achieved predictive effects that surpass humans. As the requirements for the accuracy of the model become higher and higher, the scale of model parameters and computing requirements become larger and larger, and training the model becomes a very computationally intensive and time-consuming task. Researchers often use distributed computer clusters to speed up the model training process. Distributed deep learning parallel training is dedicated to accelerating the DNN model training process, and has been studied by many scholars. Among them, the research on pipeline parallel training is getting deeper and deeper. Pipeline parallel training can solve the problem of ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/04G06N3/063G06N3/08G06F9/38
CPCG06N3/063G06N3/08G06F9/3867G06N3/045
Inventor 毛莺池屠子健聂华黄建新徐淑芳吴俊戚荣志
Owner HOHAI UNIV