Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

High-performance pipeline parallel deep neural network training

A deep neural network and training data technology, applied in the field of high-performance pipeline parallel deep neural network training, can solve problems such as the collapse of parallelization methods, and achieve the effect of reducing utilization

Pending Publication Date: 2020-12-29
MICROSOFT TECH LICENSING LLC
View PDF0 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This growth not only puts pressure on the already time- and resource-intensive DNN training process, but also breaks down common parallelization methods used to train DNNs

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • High-performance pipeline parallel deep neural network training
  • High-performance pipeline parallel deep neural network training
  • High-performance pipeline parallel deep neural network training

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] The following detailed description relates to techniques for high-performance pipelined parallel DNN model training. Among other technical benefits, the disclosed techniques can also eliminate the performance impact caused by previous parallelization techniques when training large DNN models or when network bandwidth induces a high communication-to-computation ratio. The disclosed technique can also partition the layers of a DNN model between pipeline stages to balance work and minimize network communication, and efficiently schedule forward and backward passes of a bidirectional DNN training pipeline. These and other aspects of the disclosed technology can reduce utilization of various types of computing resources, including but not limited to memory, processor cycles, network bandwidth, and power. Other technical benefits not specifically identified herein may also be realized through implementations of the disclosed technology.

[0024] Before describing the disclos...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

Layers of a deep neural network (DNN) are partitioned into stages by using a profile of the DNN. Each of the stages includes one or more of the layers of the DNN. The partitioning of the layers of theDNN into stages is optimized in various ways including optimizing the partitioning to minimize training time, to minimize data communication between worker computing devices used to train the DNN, orto ensure that the worker computing devices perform an approximately equal amount of the processing for training the DNN. The stages are assigned to the worker computing devices. The worker computingdevices process batches of training data by using a scheduling policy that causes the workers to alternate between forward processing of the batches of the DNN training data and backward processing of the batches of the DNN training data. The stages can be configured for model parallel processing or data parallel processing.

Description

Background technique [0001] In biological nervous systems such as the human brain, deep neural networks ("DNNs") are loosely modeled after information processing and communication patterns. DNNs can be used to solve complex classification problems such as but not limited to object detection, semantic labeling and feature extraction. As a result, DNNs form the basis of many artificial intelligence ("AI") applications, such as computer vision, speech recognition, and machine translation. In many domains, DNNs can match or even exceed human accuracy. [0002] The advanced performance of DNNs stems from their ability to extract advanced features from input data after using statistical learning on large datasets to obtain an efficient representation of the input space. However, the superior performance of DNNs comes at the cost of high computational complexity. High-performance general-purpose processors, such as graphics processing units ("GPUs"), are often used to provide the ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/063G06N3/08
CPCG06N3/063G06N3/08G06N3/045G06N3/04
Inventor V·塞沙德利A·费尼沙耶D·纳拉亚南A·哈拉普N·D·兰格拉詹
Owner MICROSOFT TECH LICENSING LLC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products