Unlock instant, AI-driven research and patent intelligence for your innovation.

A multi-gpu-based bpnn training method and device

A training method and technology of a training device, which are applied in the field of neural network training, can solve problems such as low efficiency and large data synchronization overhead, and achieve the effects of reducing data synchronization overhead and improving training efficiency.

Active Publication Date: 2018-04-03
BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD
View PDF2 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] However, in the above-mentioned multi-GPU BPNN training method, there is a large overhead when synchronizing the synchronization weight data between the BPNNs of the GPUs. The weight value data of a large-scale BPNN can reach hundreds of megabytes. The communication time overhead of these BPNN weight values ​​can reach hundreds of milliseconds, resulting in the inefficiency of using multiple GPUs to train BPNNs, and the training process on a single GPU usually only takes tens of milliseconds. It can be seen that due to multiple GPUs The overhead of data synchronization between them is high, resulting in low efficiency of using multiple GPUs to train BPNN, and sometimes it is not as good as using a single GPU for BPNN training

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A multi-gpu-based bpnn training method and device
  • A multi-gpu-based bpnn training method and device
  • A multi-gpu-based bpnn training method and device

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0041] figure 1 The flow chart of the multi-GPU-based BPNN training method provided by Embodiment 1 of the present invention, such as figure 1 As shown, the method includes:

[0042] S101. Control each GPU to perform forward calculation, and output O for forward calculation synchronously.

[0043] The forward calculation and reverse error calculation of BPNN are performed layer by layer, and the calculation output data of this layer can be synchronized between each GPU after the calculation of each layer is completed.

[0044]After the input layer transmits the data to the first hidden layer, each GPU is controlled to start forward calculation from the first hidden layer, and the forward calculation of each hidden layer can be completed and the forward calculation output O can be passed to the next At the same time as one hidden layer, the forward calculation output O of this layer is synchronized between each GPU, until the last layer of hidden layer transmits the forward c...

Embodiment 2

[0058] Figure 6 A schematic diagram of a multi-GPU-based BPNN training device provided in Embodiment 2 of the present invention, such as Figure 6 As shown, the device includes: a forward calculation unit 10 , a reverse error calculation unit 20 , and a weight update unit 30 .

[0059] The forward calculation unit 10 is used for controlling each GPU to perform forward calculation of BPNN, and synchronously outputting the forward calculation among each GPU.

[0060] The forward calculation and reverse error calculation of BPNN are performed layer by layer, and the calculation output data of this layer can be synchronized between each GPU after the calculation of each layer is completed.

[0061] After the data is passed from the input layer to the first hidden layer, the forward calculation unit 10 controls each GPU to start forward calculation from the first hidden layer, and the forward calculation of each hidden layer can be completed and the forward calculation While the...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The present invention provides a backpropagation neural network (BPNN) training method and device based on multiple graphics processing units (GPUs), wherein the method includes: S1, controlling each GPU to perform forward calculation of BPNN, and synchronizing between each GPU Forward calculation output; S2, controlling each GPU to perform reverse error calculation of BPNN, and synchronizing the reverse error calculation output between each GPU; S3, controlling each GPU to obtain the forward calculation output according to the synchronization and the synchronization obtained The output of the reverse error calculation updates the weights of the BPNN. The invention can reduce the data synchronization overhead during multi-GPU BPNN training, and improve the multi-GPU BPNN training efficiency.

Description

【Technical field】 [0001] The invention relates to neural network training technology, in particular to a multi-GPU-based BPNN training method and device. 【Background technique】 [0002] BPNN (Back-Propagation Nueral Networks) backpropagation neural network is a multi-layer feedforward network trained by the error backpropagation algorithm proposed by a team of scientists headed by Rumelhart and McCelland in 1986. It is currently the most widely used neural network model. one. [0003] The topology of the BPNN model includes an input layer (input), a hidden layer (hide layer) and an output layer (outputlayer). The input layer is responsible for receiving input data from the outside world and passing it to the hidden layer; the hidden layer is the internal information processing layer, responsible for data processing, and the hidden layer can be designed as a single hidden layer or a multi-hidden layer structure; the last hidden layer is passed to the output After further pr...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N3/08
Inventor 欧阳剑王勇
Owner BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD