Parallel message processing method, system and related device

A message processing and processor technology, applied in the field of data processing, can solve problems such as performance loss, and achieve the effect of reducing communication pressure

Pending Publication Date: 2020-05-05
GUANGDONG INSPUR BIG DATA RES CO LTD
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Nowadays, many mainstream deep learning frameworks use message parallel processing to generate message queues for data parallel distributed training, but

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Parallel message processing method, system and related device
  • Parallel message processing method, system and related device

Examples

Experimental program
Comparison scheme
Effect test

Example Embodiment

[0033] In order to make the purpose, technical solutions and advantages of the embodiments of the present application clearer, the following will clearly and completely describe the technical solutions in the embodiments of the present application with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments It is a part of the embodiments of this application, but not all the embodiments. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of this application.

[0034] The current Horovod deep learning framework adopts the Message passing interface specification and the message processing method of reduction to carry out distributed deep learning training. Its architecture adopts the main controller-computing structure, and the main controller acts as the message processing center to receive and send The i...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a parallel message processing method, which comprises the following steps that: when a first arithmetic unit completes the calculation of parameter units, sending a reduction operation request to a main controller, so as to enable the main controller to globally broadcast the parameter units corresponding to the reduction operation request to all arithmetic units, and enabling each arithmetic unit to cache the parameter units; after the second arithmetic unit completes calculation of the parameter units to obtain reduction data, enabling the second arithmetic unit to compare the reduction data with the parameter units in the cache of the second arithmetic unit to generate bit vector information corresponding to the second arithmetic unit; when new bit vector information is generated each time, carrying out bit AND operation on the bit vector information corresponding to each arithmetic unit; and executing reduction operation according to the bit and the operationresult. According to the application, the defect that the arithmetic unit frequently communicates and interacts with the main controller is avoided, the communication pressure on the main controlleris obviously reduced, and the bandwidth is fully utilized. The invention further provides a parallel message processing method and system, a computer readable storage medium and deep learning equipment, which have the above beneficial effects.

Description

technical field [0001] The present application relates to the field of data processing, in particular to a parallel message processing method, system and related devices. Background technique [0002] Large-scale distributed deep learning training tasks, especially for complex models and large-scale data sets, will generate a large number of message bodies and message queues during the training process. For data parallel distributed training using parallel message queue processing, multiple GPUs are independent Run the respective training data, then wait for the GPU of this node or other nodes to complete the training, and then adopt different strategies to synchronize the model parameters. In this process, a large number of message bodies and lengthy message processing waiting time are generated. The more computing units, the process takes up the longer the time. Nowadays, many mainstream deep learning frameworks use the message queue generated by the message parallel proc...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06N20/00G06F9/54
CPCG06N20/00G06F9/546
Inventor 王超
Owner GUANGDONG INSPUR BIG DATA RES CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products