Distributed parallel training method and system for neural network acoustic model

An acoustic model and neural network technology, applied to biological neural network models, speech analysis, instruments, etc., can solve problems such as extremely high network bandwidth requirements, frequent parameter transmission, and limited acceleration effects, so as to shorten the training period, prevent divergence, The effect of ensuring stability

Active Publication Date: 2017-01-04
INST OF ACOUSTICS CHINESE ACAD OF SCI +1
View PDF5 Cites 52 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] However, due to the frequent parameter transfer and extremely high requirements on network bandwidth in the process of distributed training of neural networks, most of the current parallel training systems for neural networks use one machine and insert multiple GPU cards into it. , but this method has its limitations, only four GPU cards can be used at most, and the acceleration effect that can be provided is limited

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Distributed parallel training method and system for neural network acoustic model
  • Distributed parallel training method and system for neural network acoustic model
  • Distributed parallel training method and system for neural network acoustic model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0054] 1. Construction of two-level ASGD system

[0055] attached figure 1 It is a two-level ASGD neural network system architecture diagram proposed by the present invention. The overall architecture is composed of multiple clients and a parameter server, wherein the client is responsible for calculating the gradient, and the parameter server is responsible for updating the model. Parameters are passed between servers to form an upper-level (second-level) ASGD system; each client’s internal CPU and each GPU constitute a bottom-level (first-level) ASGD system, and parameters are passed between the CPU and GPU through the bus. The process of model training based on the two-level ASGD system is as follows: first, the model in the parameter server will be initialized (random value) at the beginning of training, and the initialized model will be sent to each client (in the CPU), if each client Using 4 GPU cards (G1, G2, G3, G4), the 4 GPUs calculate the gradient according to the ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a distributed parallel training method and system for a neural network acoustic model. The distributed parallel training method comprises the steps that 101) each group of training data is inputted into a client respectively; 102) each of the client receives the inputted training data, and transmits parameters between a plurality of GPU and a first CPU, which are arranged in the client, by adopting a bus, wherein the parameters comprise the model weight and the gradient; each GPU calculates the gradient based on the inputted model weight parameter and inputs the calculated gradient into the first CPU; the first CPU updates a model copy in the corresponding client by using the gradient uploaded by the GPUs, transmits an updated weight parameter back to each GPU so as to be used for carry out gradient calculation again, and meanwhile, the first CPU accumulates the gradient inputted by each GPU and updates a model in a parameter server according to a accumulation result; and 103) latest gradient information acquired by accumulation carried out by the CPU of each client is inputted into the server by adopting network resources, and then a neural network acoustic model stored in the server is updated.

Description

technical field [0001] The invention belongs to the field of speech recognition, and is a method for using multiple computer nodes to train a neural network in parallel to improve the training speed of a neural network acoustic model, and specifically relates to a distributed parallel training method and system for a neural network acoustic model. Background technique [0002] At present, the method of using large data volume and deep neural network (Deep Neural Network, DNN) to build an acoustic model has achieved outstanding results in the field of speech recognition, which has improved the final recognition accuracy by 20% to 30%. [0003] DNN is a technology that simulates the work of neurons in the human brain by connecting weights and nodes. It can be regarded as a classifier. The DNN structure mainly includes an input layer, a hidden layer and an output layer. There are weighted line connections, the number of nodes in the output layer is determined by the number of t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G10L15/06G10L15/16G10L15/30G06N3/02
Inventor 那兴宇王智超潘接林颜永红
Owner INST OF ACOUSTICS CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products