Supercharge Your Innovation With Domain-Expert AI Agents!

Asynchronous training of machine learning models

A technology of machine learning model and training data, applied in the field of asynchronous training of machine learning model, which can solve problems such as mismatch of working machines

Active Publication Date: 2021-06-15
MICROSOFT TECH LICENSING LLC
View PDF9 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the core problem of distributed or asynchronous model training is the mismatch between individual workers

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Asynchronous training of machine learning models
  • Asynchronous training of machine learning models
  • Asynchronous training of machine learning models

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0012] The present disclosure will now be discussed with reference to several embodiments. It should be understood that these embodiments are discussed only to enable those of ordinary skill in the art to better understand and thus implement the present disclosure, and do not imply any limitation on the scope of the present disclosure.

[0013] As used herein, the term "comprising" and variations thereof are to be read as open-ended terms meaning "including but not limited to". The term "based on" is to be read as "based at least in part on". The terms "an embodiment" and "one embodiment" are to be read as "at least one embodiment". The term "another embodiment" is to be read as "at least one other embodiment". The terms "first", "second", etc. may refer to different or the same object. Other definitions, both express and implied, may also be included below.

[0014] Asynchronous Training Architecture

[0015] figure 1 A block diagram of a parallel computing environment ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

Embodiments of the present disclosure relate to asynchronous training of machine learning models. The server receives feedback data generated by training the machine learning model from the worker machine. These feedback data are obtained by the worker using its own training data and are associated with the previous values ​​of the parameter sets of the machine learning model at this particular worker. The server determines a difference between the previous value and a current value of the parameter set at the server. This current value may have been updated one or more times due to the actions of other workers. The server can then update the current value of the parameter set based on the difference between the feedback data and the value of the parameter set. Thus, this update not only takes into account the training results of each worker, but also properly compensates for the delay between different workers.

Description

Background technique [0001] Machine learning has wide-ranging applications in fields such as speech recognition, computer vision, and natural language processing. For example, Deep Neural Networks (DNN) can train machine learning models with multiple layers and parameters in parallel on the basis of big data and powerful computing resources. In the training phase, one or more parameters of the model need to be trained according to the given training data set and optimization goal. For example, for the training of neural networks, stochastic gradient descent methods can be used. [0002] It is known that the training dataset can be distributed among multiple workers. These worker machines use their respective training data to optimize model parameters and return their results to a central server. However, the core problem of distributed or asynchronous model training is the mismatch between individual workers. For example, when a worker returns its parameter updates, the mo...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06N20/00G06N3/02
CPCG06N20/00G06N3/02G06N3/084G06N3/045G06N3/04G06N3/08
Inventor 王太峰陈薇刘铁岩高飞叶启威
Owner MICROSOFT TECH LICENSING LLC
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More