Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Training method and device for acceleration distributed deep neural network

A deep neural network, distributed technology, applied in neural learning methods, biological neural network models, neural architectures, etc., can solve problems such as long training time of deep neural networks

Inactive Publication Date: 2018-11-23
BEIJING UNIV OF POSTS & TELECOMM
View PDF5 Cites 27 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, during the training process, a serious problem also occurred. With the explosive growth of network parameters and training samples, the training time of deep neural network was very long.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Training method and device for acceleration distributed deep neural network
  • Training method and device for acceleration distributed deep neural network
  • Training method and device for acceleration distributed deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment approach

[0164] As an implementation manner of the embodiment of the present invention, the above-mentioned model mapping unit may include:

[0165] The model mapping subunit ( Figure 5 Not shown in ), used to map the scheduling of the multiple tasks into the following directed graph model:

[0166] The directed graph model includes source nodes, application nodes, task nodes, computing nodes, and destination nodes, as well as directed edges from source nodes to application nodes, directed edges from application nodes to task nodes, and directed edges from task nodes to computing nodes. Edges, directed edges from computing nodes to destination nodes;

[0167] Wherein, each node object has a potential value, and the node object includes the source node, the application node, the task node, the computing node and the destination node; the potential value is positive, indicating that the node object has Tasks can be assigned, and the number of tasks that can be assigned is the potentia...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention provides a training method and device for an acceleration distributed deep neural network. The method comprises the following steps that: on the basis of parallel training, designing the training of the deep neural network into a distributed training pattern, and dividing a deep neural network model to be trained into a plurality of sub-networks; dividing a trainingsample set into a plurality of sub-sample sets; and on the basis of distributed cluster architecture and a preset scheduling method, utilizing the plurality of sub-sample sets to train the deep neural network, wherein each piece of training is carried out by a plurality of sub-networks so as to finish the distributed training of the deep neural network. Since the influence of network delay for the distributed training sub-network can be reduced through data localization on the basis of the distributed cluster architecture and the preset scheduling method, a training strategy is regulated in real time and the progresses of parallel-training sub-networks can be synchronized, the training completion time of the distributed deep neural network can be shortened, and the training of the deep neural network is accelerated.

Description

technical field [0001] The invention relates to the technical field of deep neural network training, in particular to a training method and device for accelerating a distributed deep neural network. Background technique [0002] Deep neural networks have been successfully applied in many fields, including image recognition, texture classification, speech recognition and other fields. In recent years, the performance of deep neural networks has been significantly improved due to the use of deeper network architectures and larger training sample sets for training. However, in the training process, a serious problem also arises. With the explosive growth of network parameters and training samples, the training time of deep neural network is very long. [0003] In order to solve this problem, researchers have proposed a method for parallel training of deep neural networks, mainly using multiple graphics processor cores to achieve parallel training, thereby reducing training tim...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T1/20G06N3/10G06N3/08
CPCG06N3/084G06N3/10G06T1/20G06N3/08G06F9/4881G06N3/045G06N3/04
Inventor 廖建新王敬宇王晶戚琦徐捷
Owner BEIJING UNIV OF POSTS & TELECOMM
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products